diff --git a/removed_sites.json b/removed_sites.json index 2dbee82..bc0cba9 100644 --- a/removed_sites.json +++ b/removed_sites.json @@ -837,5 +837,37 @@ "url": "https://forums.gunsandammo.com/profile/{}", "urlMain": "https://gunsandammo.com/", "username_claimed": "adam" + }, + "TikTok": { + "errorType": "status_code", + "url": "https://tiktok.com/@{}", + "urlMain": "https://tiktok.com/", + "username_claimed": "red" + }, + "Lolchess": { + "errorMsg": "No search results", + "errorType": "message", + "url": "https://lolchess.gg/profile/na/{}", + "urlMain": "https://lolchess.gg/", + "username_claimed": "blue" + }, + "Virgool": { + "errorMsg": "\u06f4\u06f0\u06f4", + "errorType": "message", + "url": "https://virgool.io/@{}", + "urlMain": "https://virgool.io/", + "username_claimed": "blue" + }, + "Whonix Forum": { + "errorType": "status_code", + "url": "https://forums.whonix.org/u/{}/summary", + "urlMain": "https://forums.whonix.org/", + "username_claimed": "red" + }, + "ebio.gg": { + "errorType": "status_code", + "url": "https://ebio.gg/{}", + "urlMain": "https:/ebio.gg", + "username_claimed": "dev" } } diff --git a/removed_sites.md b/removed_sites.md index 2a23e6a..b1e5aa5 100644 --- a/removed_sites.md +++ b/removed_sites.md @@ -1824,4 +1824,62 @@ As of 2023.08.29, GunsAndAmmo responds with 404 from time to time "urlMain": "https://gunsandammo.com/", "username_claimed": "adam" } +``` + +## TikTok +As of 2023.12.21, TikTok returns false positives. This is because the webpage returns a somewhat blank page. This prevents us from being able to check for the existance of usernames. Proxitok does not work either. + +```json + "TikTok": { + "errorType": "status_code", + "url": "https://tiktok.com/@{}", + "urlMain": "https://tiktok.com/", + "username_claimed": "red" + }, +``` + +## Lolchess +As of 2023.12.21, Lolchess returns false positives. +```json + "Lolchess": { + "errorMsg": "No search results", + "errorType": "message", + "url": "https://lolchess.gg/profile/na/{}", + "urlMain": "https://lolchess.gg/", + "username_claimed": "blue" + }, +``` + +## Virgool +As of 2023.12.21, Lolchess returns false positives. +```json + "Virgool": { + "errorMsg": "\u06f4\u06f0\u06f4", + "errorType": "message", + "url": "https://virgool.io/@{}", + "urlMain": "https://virgool.io/", + "username_claimed": "blue" + }, +``` + +## Whonix Forum +As of 2023.12.21, Lolchess returns false positives. +```json + "Whonix Forum": { + "errorType": "status_code", + "url": "https://forums.whonix.org/u/{}/summary", + "urlMain": "https://forums.whonix.org/", + "username_claimed": "red" + }, +``` + +## Ebio +As of 2023.12.21, Lolchess returns false positives. +```json + "ebio.gg": { + "errorType": "status_code", + "url": "https://ebio.gg/{}", + "urlMain": "https:/ebio.gg", + "username_claimed": "dev" + }, ``` \ No newline at end of file diff --git a/sherlock/resources/data.json b/sherlock/resources/data.json index 6a12ecd..d45b806 100644 --- a/sherlock/resources/data.json +++ b/sherlock/resources/data.json @@ -502,7 +502,7 @@ "username_claimed": "blue" }, "Contently": { - "errorMsg": "We can't find that page!", + "errorMsg": "Leading Content Marketing Platform | Contently", "errorType": "message", "regexCheck": "^[a-zA-Z][a-zA-Z0-9_-]*$", "url": "https://{}.contently.com/", @@ -1237,13 +1237,6 @@ "urlMain": "https://lobste.rs/", "username_claimed": "jcs" }, - "Lolchess": { - "errorMsg": "No search results", - "errorType": "message", - "url": "https://lolchess.gg/profile/na/{}", - "urlMain": "https://lolchess.gg/", - "username_claimed": "blue" - }, "LottieFiles": { "errorType": "status_code", "url": "https://lottiefiles.com/{}", @@ -1945,12 +1938,6 @@ "urlMain": "https://themeforest.net/", "username_claimed": "user" }, - "TikTok": { - "errorType": "status_code", - "url": "https://tiktok.com/@{}", - "urlMain": "https://tiktok.com/", - "username_claimed": "red" - }, "TnAFlix": { "errorType": "status_code", "isNSFW": true, @@ -2097,13 +2084,6 @@ "urlMain": "https://vimeo.com/", "username_claimed": "blue" }, - "Virgool": { - "errorMsg": "\u06f4\u06f0\u06f4", - "errorType": "message", - "url": "https://virgool.io/@{}", - "urlMain": "https://virgool.io/", - "username_claimed": "blue" - }, "VirusTotal": { "errorType": "status_code", "request_method": "GET", @@ -2152,12 +2132,6 @@ "urlMain": "https://weebly.com/", "username_claimed": "blue" }, - "Whonix Forum": { - "errorType": "status_code", - "url": "https://forums.whonix.org/u/{}/summary", - "urlMain": "https://forums.whonix.org/", - "username_claimed": "red" - }, "Wikidot": { "errorMsg": "User does not exist.", "errorType": "message", @@ -2344,12 +2318,6 @@ "urlMain": "https://egpu.io/", "username_claimed": "blue" }, - "ebio.gg": { - "errorType": "status_code", - "url": "https://ebio.gg/{}", - "urlMain": "https:/ebio.gg", - "username_claimed": "dev" - }, "eintracht": { "errorType": "status_code", "url": "https://community.eintracht.de/fans/{}", diff --git a/sherlock/sherlock.py b/sherlock/sherlock.py index 7256195..bf68689 100644 --- a/sherlock/sherlock.py +++ b/sherlock/sherlock.py @@ -26,6 +26,7 @@ from result import QueryResult from notify import QueryNotifyPrint from sites import SitesInformation from colorama import init +from argparse import ArgumentTypeError module_name = "Sherlock: Find Usernames Across Social Networks" __version__ = "0.14.3" @@ -91,10 +92,9 @@ class SherlockFuturesSession(FuturesSession): # No response hook was already defined, so install it ourselves. hooks["response"] = [response_time] - return super(SherlockFuturesSession, self).request(method, - url, - hooks=hooks, - *args, **kwargs) + return super(SherlockFuturesSession, self).request( + method, url, hooks=hooks, *args, **kwargs + ) def get_response(request_future, error_type, social_network): @@ -127,42 +127,43 @@ def get_response(request_future, error_type, social_network): return response, error_context, exception_text -def interpolate_string(object, username): - """Insert a string into the string properties of an object recursively.""" - - if isinstance(object, str): - return object.replace("{}", username) - elif isinstance(object, dict): - for key, value in object.items(): - object[key] = interpolate_string(value, username) - elif isinstance(object, list): - for i in object: - object[i] = interpolate_string(object[i], username) - - return object +def interpolate_string(input_object, username): + if isinstance(input_object, str): + return input_object.replace("{}", username) + elif isinstance(input_object, dict): + return {k: interpolate_string(v, username) for k, v in input_object.items()} + elif isinstance(input_object, list): + return [interpolate_string(i, username) for i in input_object] + return input_object -def CheckForParameter(username): - '''checks if {?} exists in the username - if exist it means that sherlock is looking for more multiple username''' - return ("{?}" in username) +def check_for_parameter(username): + """checks if {?} exists in the username + if exist it means that sherlock is looking for more multiple username""" + return "{?}" in username checksymbols = [] checksymbols = ["_", "-", "."] -def MultipleUsernames(username): - '''replace the parameter with with symbols and return a list of usernames''' +def multiple_usernames(username): + """replace the parameter with with symbols and return a list of usernames""" allUsernames = [] for i in checksymbols: allUsernames.append(username.replace("{?}", i)) return allUsernames -def sherlock(username, site_data, query_notify, - tor=False, unique_tor=False, - proxy=None, timeout=60): +def sherlock( + username, + site_data, + query_notify, + tor=False, + unique_tor=False, + proxy=None, + timeout=60, +): """Run Sherlock Analysis. Checks for existence of username on various social media sites. @@ -214,15 +215,15 @@ def sherlock(username, site_data, query_notify, max_workers = len(site_data) # Create multi-threaded session for all requests. - session = SherlockFuturesSession(max_workers=max_workers, - session=underlying_session) + session = SherlockFuturesSession( + max_workers=max_workers, session=underlying_session + ) # Results from analysis of all sites results_total = {} # First create futures for all requests. This allows for the requests to run in parallel for social_network, net_info in site_data.items(): - # Results from analysis of this specific site results_site = {"url_main": net_info.get("urlMain")} @@ -245,10 +246,9 @@ def sherlock(username, site_data, query_notify, regex_check = net_info.get("regexCheck") if regex_check and re.search(regex_check, username) is None: # No need to do the check at the site: this username is not allowed. - results_site["status"] = QueryResult(username, - social_network, - url, - QueryStatus.ILLEGAL) + results_site["status"] = QueryResult( + username, social_network, url, QueryStatus.ILLEGAL + ) results_site["url_user"] = "" results_site["http_status"] = "" results_site["response_text"] = "" @@ -309,18 +309,22 @@ def sherlock(username, site_data, query_notify, # This future starts running the request in a new thread, doesn't block the main thread if proxy is not None: proxies = {"http": proxy, "https": proxy} - future = request(url=url_probe, headers=headers, - proxies=proxies, - allow_redirects=allow_redirects, - timeout=timeout, - json=request_payload - ) + future = request( + url=url_probe, + headers=headers, + proxies=proxies, + allow_redirects=allow_redirects, + timeout=timeout, + json=request_payload, + ) else: - future = request(url=url_probe, headers=headers, - allow_redirects=allow_redirects, - timeout=timeout, - json=request_payload - ) + future = request( + url=url_probe, + headers=headers, + allow_redirects=allow_redirects, + timeout=timeout, + json=request_payload, + ) # Store future in data for access later net_info["request_future"] = future @@ -335,7 +339,6 @@ def sherlock(username, site_data, query_notify, # Open the file containing account links # Core logic: If tor requests, make them here. If multi-threaded requests, wait for responses for social_network, net_info in site_data.items(): - # Retrieve results again results_site = results_total.get(social_network) @@ -352,9 +355,9 @@ def sherlock(username, site_data, query_notify, # Retrieve future and ensure it has finished future = net_info["request_future"] - r, error_text, exception_text = get_response(request_future=future, - error_type=error_type, - social_network=social_network) + r, error_text, exception_text = get_response( + request_future=future, error_type=error_type, social_network=social_network + ) # Get response time for response of our request. try: @@ -365,11 +368,11 @@ def sherlock(username, site_data, query_notify, # Attempt to get request information try: http_status = r.status_code - except: + except Exception: http_status = "?" try: response_text = r.text.encode(r.encoding or "UTF-8") - except: + except Exception: response_text = "" query_status = QueryStatus.UNKNOWN @@ -424,16 +427,19 @@ def sherlock(username, site_data, query_notify, query_status = QueryStatus.AVAILABLE else: # It should be impossible to ever get here... - raise ValueError(f"Unknown Error Type '{error_type}' for " - f"site '{social_network}'") + raise ValueError( + f"Unknown Error Type '{error_type}' for " f"site '{social_network}'" + ) # Notify caller about results of query. - result = QueryResult(username=username, - site_name=social_network, - site_url_user=url, - status=query_status, - query_time=response_time, - context=error_context) + result = QueryResult( + username=username, + site_name=social_network, + site_url_user=url, + status=query_status, + query_time=response_time, + context=error_context, + ) query_notify.update(result) # Save status of request @@ -463,16 +469,13 @@ def timeout_check(value): NOTE: Will raise an exception if the timeout in invalid. """ - from argparse import ArgumentTypeError - try: - timeout = float(value) - except: - raise ArgumentTypeError(f"Timeout '{value}' must be a number.") - if timeout <= 0: + if value <= 0: raise ArgumentTypeError( - f"Timeout '{value}' must be greater than 0.0s.") - return timeout + f"Invalid timeout value: {value}. Timeout must be a positive number." + ) + + return float(value) def handler(signal_received, frame): @@ -484,86 +487,159 @@ def handler(signal_received, frame): def main(): - version_string = f"%(prog)s {__version__}\n" + \ - f"{requests.__description__}: {requests.__version__}\n" + \ - f"Python: {platform.python_version()}" + version_string = ( + f"%(prog)s {__version__}\n" + + f"{requests.__description__}: {requests.__version__}\n" + + f"Python: {platform.python_version()}" + ) - parser = ArgumentParser(formatter_class=RawDescriptionHelpFormatter, - description=f"{module_name} (Version {__version__})" - ) - parser.add_argument("--version", - action="version", version=version_string, - help="Display version information and dependencies." - ) - parser.add_argument("--verbose", "-v", "-d", "--debug", - action="store_true", dest="verbose", default=False, - help="Display extra debugging information and metrics." - ) - parser.add_argument("--folderoutput", "-fo", dest="folderoutput", - help="If using multiple usernames, the output of the results will be saved to this folder." - ) - parser.add_argument("--output", "-o", dest="output", - help="If using single username, the output of the result will be saved to this file." - ) - parser.add_argument("--tor", "-t", - action="store_true", dest="tor", default=False, - help="Make requests over Tor; increases runtime; requires Tor to be installed and in system path.") - parser.add_argument("--unique-tor", "-u", - action="store_true", dest="unique_tor", default=False, - help="Make requests over Tor with new Tor circuit after each request; increases runtime; requires Tor to be installed and in system path.") - parser.add_argument("--csv", - action="store_true", dest="csv", default=False, - help="Create Comma-Separated Values (CSV) File." - ) - parser.add_argument("--xlsx", - action="store_true", dest="xlsx", default=False, - help="Create the standard file for the modern Microsoft Excel spreadsheet (xslx)." - ) - parser.add_argument("--site", - action="append", metavar="SITE_NAME", - dest="site_list", default=None, - help="Limit analysis to just the listed sites. Add multiple options to specify more than one site." - ) - parser.add_argument("--proxy", "-p", metavar="PROXY_URL", - action="store", dest="proxy", default=None, - help="Make requests over a proxy. e.g. socks5://127.0.0.1:1080" - ) - parser.add_argument("--json", "-j", metavar="JSON_FILE", - dest="json_file", default=None, - help="Load data from a JSON file or an online, valid, JSON file.") - parser.add_argument("--timeout", - action="store", metavar="TIMEOUT", - dest="timeout", type=timeout_check, default=60, - help="Time (in seconds) to wait for response to requests (Default: 60)" - ) - parser.add_argument("--print-all", - action="store_true", dest="print_all", default=False, - help="Output sites where the username was not found." - ) - parser.add_argument("--print-found", - action="store_true", dest="print_found", default=True, - help="Output sites where the username was found (also if exported as file)." - ) - parser.add_argument("--no-color", - action="store_true", dest="no_color", default=False, - help="Don't color terminal output" - ) - parser.add_argument("username", - nargs="+", metavar="USERNAMES", - action="store", - help="One or more usernames to check with social networks. Check similar usernames using {%%} (replace to '_', '-', '.')." - ) - parser.add_argument("--browse", "-b", - action="store_true", dest="browse", default=False, - help="Browse to all results on default browser.") + parser = ArgumentParser( + formatter_class=RawDescriptionHelpFormatter, + description=f"{module_name} (Version {__version__})", + ) + parser.add_argument( + "--version", + action="version", + version=version_string, + help="Display version information and dependencies.", + ) + parser.add_argument( + "--verbose", + "-v", + "-d", + "--debug", + action="store_true", + dest="verbose", + default=False, + help="Display extra debugging information and metrics.", + ) + parser.add_argument( + "--folderoutput", + "-fo", + dest="folderoutput", + help="If using multiple usernames, the output of the results will be saved to this folder.", + ) + parser.add_argument( + "--output", + "-o", + dest="output", + help="If using single username, the output of the result will be saved to this file.", + ) + parser.add_argument( + "--tor", + "-t", + action="store_true", + dest="tor", + default=False, + help="Make requests over Tor; increases runtime; requires Tor to be installed and in system path.", + ) + parser.add_argument( + "--unique-tor", + "-u", + action="store_true", + dest="unique_tor", + default=False, + help="Make requests over Tor with new Tor circuit after each request; increases runtime; requires Tor to be installed and in system path.", + ) + parser.add_argument( + "--csv", + action="store_true", + dest="csv", + default=False, + help="Create Comma-Separated Values (CSV) File.", + ) + parser.add_argument( + "--xlsx", + action="store_true", + dest="xlsx", + default=False, + help="Create the standard file for the modern Microsoft Excel spreadsheet (xslx).", + ) + parser.add_argument( + "--site", + action="append", + metavar="SITE_NAME", + dest="site_list", + default=None, + help="Limit analysis to just the listed sites. Add multiple options to specify more than one site.", + ) + parser.add_argument( + "--proxy", + "-p", + metavar="PROXY_URL", + action="store", + dest="proxy", + default=None, + help="Make requests over a proxy. e.g. socks5://127.0.0.1:1080", + ) + parser.add_argument( + "--json", + "-j", + metavar="JSON_FILE", + dest="json_file", + default=None, + help="Load data from a JSON file or an online, valid, JSON file.", + ) + parser.add_argument( + "--timeout", + action="store", + metavar="TIMEOUT", + dest="timeout", + type=timeout_check, + default=60, + help="Time (in seconds) to wait for response to requests (Default: 60)", + ) + parser.add_argument( + "--print-all", + action="store_true", + dest="print_all", + default=False, + help="Output sites where the username was not found.", + ) + parser.add_argument( + "--print-found", + action="store_true", + dest="print_found", + default=True, + help="Output sites where the username was found (also if exported as file).", + ) + parser.add_argument( + "--no-color", + action="store_true", + dest="no_color", + default=False, + help="Don't color terminal output", + ) + parser.add_argument( + "username", + nargs="+", + metavar="USERNAMES", + action="store", + help="One or more usernames to check with social networks. Check similar usernames using {%%} (replace to '_', '-', '.').", + ) + parser.add_argument( + "--browse", + "-b", + action="store_true", + dest="browse", + default=False, + help="Browse to all results on default browser.", + ) - parser.add_argument("--local", "-l", - action="store_true", default=False, - help="Force the use of the local data.json file.") + parser.add_argument( + "--local", + "-l", + action="store_true", + default=False, + help="Force the use of the local data.json file.", + ) - parser.add_argument("--nsfw", - action="store_true", default=False, - help="Include checking of NSFW sites from default list.") + parser.add_argument( + "--nsfw", + action="store_true", + default=False, + help="Include checking of NSFW sites from default list.", + ) args = parser.parse_args() @@ -573,14 +649,17 @@ def main(): # Check for newer version of Sherlock. If it exists, let the user know about it try: r = requests.get( - "https://raw.githubusercontent.com/sherlock-project/sherlock/master/sherlock/sherlock.py") + "https://raw.githubusercontent.com/sherlock-project/sherlock/master/sherlock/sherlock.py" + ) remote_version = str(re.findall('__version__ = "(.*)"', r.text)[0]) local_version = __version__ if remote_version != local_version: - print("Update Available!\n" + - f"You are running version {local_version}. Version {remote_version} is available at https://github.com/sherlock-project/sherlock") + print( + "Update Available!\n" + + f"You are running version {local_version}. Version {remote_version} is available at https://github.com/sherlock-project/sherlock" + ) except Exception as error: print(f"A problem occurred while checking for an update: {error}") @@ -598,7 +677,8 @@ def main(): print("Using Tor to make requests") print( - "Warning: some websites might refuse connecting over Tor, so note that using this option might increase connection errors.") + "Warning: some websites might refuse connecting over Tor, so note that using this option might increase connection errors." + ) if args.no_color: # Disable color output. @@ -620,8 +700,9 @@ def main(): # Create object with all information about sites we are aware of. try: if args.local: - sites = SitesInformation(os.path.join( - os.path.dirname(__file__), "resources/data.json")) + sites = SitesInformation( + os.path.join(os.path.dirname(__file__), "resources/data.json") + ) else: sites = SitesInformation(args.json_file) except Exception as error: @@ -654,35 +735,34 @@ def main(): site_missing.append(f"'{site}'") if site_missing: - print( - f"Error: Desired sites not found: {', '.join(site_missing)}.") + print(f"Error: Desired sites not found: {', '.join(site_missing)}.") if not site_data: sys.exit(1) # Create notify object for query results. - query_notify = QueryNotifyPrint(result=None, - verbose=args.verbose, - print_all=args.print_all, - browse=args.browse) + query_notify = QueryNotifyPrint( + result=None, verbose=args.verbose, print_all=args.print_all, browse=args.browse + ) # Run report on all specified users. all_usernames = [] for username in args.username: - if (CheckForParameter(username)): - for name in MultipleUsernames(username): + if check_for_parameter(username): + for name in multiple_usernames(username): all_usernames.append(name) else: all_usernames.append(username) for username in all_usernames: - - results = sherlock(username, - site_data, - query_notify, - tor=args.tor, - unique_tor=args.unique_tor, - proxy=args.proxy, - timeout=args.timeout) + results = sherlock( + username, + site_data, + query_notify, + tor=args.tor, + unique_tor=args.unique_tor, + proxy=args.proxy, + timeout=args.timeout, + ) if args.output: result_file = args.output @@ -701,8 +781,7 @@ def main(): if dictionary.get("status").status == QueryStatus.CLAIMED: exists_counter += 1 file.write(dictionary["url_user"] + "\n") - file.write( - f"Total Websites Username Detected On : {exists_counter}\n") + file.write(f"Total Websites Username Detected On : {exists_counter}\n") if args.csv: result_file = f"{username}.csv" @@ -712,33 +791,41 @@ def main(): os.makedirs(args.folderoutput, exist_ok=True) result_file = os.path.join(args.folderoutput, result_file) - with open(result_file, "w", newline='', encoding="utf-8") as csv_report: + with open(result_file, "w", newline="", encoding="utf-8") as csv_report: writer = csv.writer(csv_report) - writer.writerow(["username", - "name", - "url_main", - "url_user", - "exists", - "http_status", - "response_time_s" - ] - ) + writer.writerow( + [ + "username", + "name", + "url_main", + "url_user", + "exists", + "http_status", + "response_time_s", + ] + ) for site in results: - if args.print_found and not args.print_all and results[site]["status"].status != QueryStatus.CLAIMED: + if ( + args.print_found + and not args.print_all + and results[site]["status"].status != QueryStatus.CLAIMED + ): continue response_time_s = results[site]["status"].query_time if response_time_s is None: response_time_s = "" - writer.writerow([username, - site, - results[site]["url_main"], - results[site]["url_user"], - str(results[site]["status"].status), - results[site]["http_status"], - response_time_s - ] - ) + writer.writerow( + [ + username, + site, + results[site]["url_main"], + results[site]["url_user"], + str(results[site]["status"].status), + results[site]["http_status"], + response_time_s, + ] + ) if args.xlsx: usernames = [] names = [] @@ -749,7 +836,11 @@ def main(): response_time_s = [] for site in results: - if args.print_found and not args.print_all and results[site]["status"].status != QueryStatus.CLAIMED: + if ( + args.print_found + and not args.print_all + and results[site]["status"].status != QueryStatus.CLAIMED + ): continue if response_time_s is None: @@ -763,8 +854,18 @@ def main(): exists.append(str(results[site]["status"].status)) http_status.append(results[site]["http_status"]) - DataFrame = pd.DataFrame({"username": usernames, "name": names, "url_main": url_main, "url_user": url_user, "exists": exists, "http_status": http_status, "response_time_s": response_time_s}) - DataFrame.to_excel(f'{username}.xlsx', sheet_name='sheet1', index=False) + DataFrame = pd.DataFrame( + { + "username": usernames, + "name": names, + "url_main": url_main, + "url_user": url_user, + "exists": exists, + "http_status": http_status, + "response_time_s": response_time_s, + } + ) + DataFrame.to_excel(f"{username}.xlsx", sheet_name="sheet1", index=False) print() query_notify.finish() diff --git a/sherlock/tests/test_multiple_usernames.py b/sherlock/tests/test_multiple_usernames.py index 0692e8f..4181108 100644 --- a/sherlock/tests/test_multiple_usernames.py +++ b/sherlock/tests/test_multiple_usernames.py @@ -1,4 +1,4 @@ -import imp +import importlib import unittest import sys sys.path.append('../') @@ -7,9 +7,9 @@ import sherlock as sh checksymbols = [] checksymbols = ["_", "-", "."] -"""Test for mulriple usernames. +"""Test for multiple usernames. - This test ensures that the function MultipleUsernames works properly. More specific, + This test ensures that the function multiple_usernames works properly. More specific, different scenarios are tested and only usernames that contain this specific sequence: {?} should return positive. @@ -23,7 +23,7 @@ class TestMultipleUsernames(unittest.TestCase): def test_area(self): test_usernames = ["test{?}test" , "test{?feo" , "test"] for name in test_usernames: - if(sh.CheckForParameter(name)): - self.assertAlmostEqual(sh.MultipleUsernames(name), ["test_test" , "test-test" , "test.test"]) + if(sh.check_for_parameter(name)): + self.assertAlmostEqual(sh.multiple_usernames(name), ["test_test" , "test-test" , "test.test"]) else: self.assertAlmostEqual(name, name) \ No newline at end of file diff --git a/sites.md b/sites.md index 96bd0d9..4d07404 100644 --- a/sites.md +++ b/sites.md @@ -1,4 +1,4 @@ -## List Of Supported Sites (401 Sites In Total!) +## List Of Supported Sites (396 Sites In Total!) 1. ![](https://www.google.com/s2/favicons?domain=https://2Dimensions.com/) [2Dimensions](https://2Dimensions.com/) 1. ![](https://www.google.com/s2/favicons?domain=http://forum.3dnews.ru/) [3dnews](http://forum.3dnews.ru/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.7cups.com/) [7Cups](https://www.7cups.com/) @@ -180,7 +180,6 @@ 1. ![](https://www.google.com/s2/favicons?domain=https://listed.to/) [Listed](https://listed.to/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.livejournal.com/) [LiveJournal](https://www.livejournal.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://lobste.rs/) [Lobsters](https://lobste.rs/) -1. ![](https://www.google.com/s2/favicons?domain=https://lolchess.gg/) [Lolchess](https://lolchess.gg/) 1. ![](https://www.google.com/s2/favicons?domain=https://lottiefiles.com/) [LottieFiles](https://lottiefiles.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.lushstories.com/) [LushStories](https://www.lushstories.com/) **(NSFW)** 1. ![](https://www.google.com/s2/favicons?domain=https://forums.mmorpg.com/) [MMORPG Forum](https://forums.mmorpg.com/) @@ -286,7 +285,6 @@ 1. ![](https://www.google.com/s2/favicons?domain=https://tellonym.me/) [Tellonym.me](https://tellonym.me/) 1. ![](https://www.google.com/s2/favicons?domain=https://tenor.com/) [Tenor](https://tenor.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://themeforest.net/) [ThemeForest](https://themeforest.net/) -1. ![](https://www.google.com/s2/favicons?domain=https://tiktok.com/) [TikTok](https://tiktok.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.tnaflix.com/) [TnAFlix](https://www.tnaflix.com/) **(NSFW)** 1. ![](https://www.google.com/s2/favicons?domain=https://www.tradingview.com/) [TradingView](https://www.tradingview.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.trakt.tv/) [Trakt](https://www.trakt.tv/) @@ -307,7 +305,6 @@ 1. ![](https://www.google.com/s2/favicons?domain=https://venmo.com/) [Venmo](https://venmo.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://vero.co/) [Vero](https://vero.co/) 1. ![](https://www.google.com/s2/favicons?domain=https://vimeo.com/) [Vimeo](https://vimeo.com/) -1. ![](https://www.google.com/s2/favicons?domain=https://virgool.io/) [Virgool](https://virgool.io/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.virustotal.com/) [VirusTotal](https://www.virustotal.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://discourse.wicg.io/) [WICG Forum](https://discourse.wicg.io/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.warriorforum.com/) [Warrior Forum](https://www.warriorforum.com/) @@ -315,7 +312,6 @@ 1. ![](https://www.google.com/s2/favicons?domain=https://www.webnode.cz/) [WebNode](https://www.webnode.cz/) 1. ![](https://www.google.com/s2/favicons?domain=https://hosted.weblate.org/) [Weblate](https://hosted.weblate.org/) 1. ![](https://www.google.com/s2/favicons?domain=https://weebly.com/) [Weebly](https://weebly.com/) -1. ![](https://www.google.com/s2/favicons?domain=https://forums.whonix.org/) [Whonix Forum](https://forums.whonix.org/) 1. ![](https://www.google.com/s2/favicons?domain=http://www.wikidot.com/) [Wikidot](http://www.wikidot.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.wikipedia.org/) [Wikipedia](https://www.wikipedia.org/) 1. ![](https://www.google.com/s2/favicons?domain=https://windy.com/) [Windy](https://windy.com/) @@ -344,7 +340,6 @@ 1. ![](https://www.google.com/s2/favicons?domain=https://devrant.com/) [devRant](https://devrant.com/) 1. ![](https://www.google.com/s2/favicons?domain=https://www.drive2.ru/) [drive2](https://www.drive2.ru/) 1. ![](https://www.google.com/s2/favicons?domain=https://egpu.io/) [eGPU](https://egpu.io/) -1. ![](https://www.google.com/s2/favicons?domain=https:/ebio.gg) [ebio.gg](https:/ebio.gg) 1. ![](https://www.google.com/s2/favicons?domain=https://eintracht.de) [eintracht](https://eintracht.de) 1. ![](https://www.google.com/s2/favicons?domain=https://www.fixya.com) [fixya](https://www.fixya.com) 1. ![](https://www.google.com/s2/favicons?domain=https://www.fl.ru/) [fl](https://www.fl.ru/)