mirror of
https://bitbucket.org/anticensority/antizapret-pac-generator-light.git
synced 2024-11-23 21:56:04 +03:00
165 lines
5.6 KiB
Python
Executable File
165 lines
5.6 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
|
|
import sys
|
|
import os
|
|
import asyncio
|
|
import dns.resolver
|
|
import dns.rdatatype
|
|
import dns.asyncresolver
|
|
import dns.exception
|
|
import dns._asyncio_backend
|
|
import itertools
|
|
|
|
# DNS timeout (in seconds) for the initial DNS resolving pass
|
|
INITIAL_PASS_TIMEOUT = 3
|
|
# Number of concurrent resolving 'threads' for initial pass
|
|
INITIAL_PASS_CONCURRENCY = 100
|
|
|
|
# DNS timeout (in seconds) for the final (second) DNS resolving pass
|
|
FINAL_PASS_TIMEOUT = 10
|
|
# Number of concurrent resolving 'threads' for final pass
|
|
FINAL_PASS_CONCURRENCY = 35
|
|
|
|
# Multiplication number of asynchronous tasks to build for a single
|
|
# resolver loop
|
|
TASK_AMOUNT_MULTIPLIER = 5
|
|
|
|
# Parked and expired domains NS record substrings
|
|
NS_FILTER_SUBSTRINGS = ("parking", "expired", ".afternic.com", "parklogic", ".parktons.com",
|
|
".above.com", ".ztomy.com", ".notneiron.com", ".ibspark.com", ".bodis.com",
|
|
".example.com")
|
|
|
|
class AZResolver(dns.asyncresolver.Resolver):
|
|
def __init__(self, *args, **kwargs):
|
|
self.limitConcurrency(25) # default limit
|
|
super().__init__(*args, **kwargs)
|
|
|
|
def limitConcurrency(self, count):
|
|
self.limitingsemaphore = asyncio.Semaphore(count)
|
|
|
|
async def nxresolve(self, domain):
|
|
async with self.limitingsemaphore:
|
|
try:
|
|
#print(domain, file=sys.stderr)
|
|
domain_nses = await self.resolve(domain, dns.rdatatype.NS)
|
|
for ns in domain_nses:
|
|
# Filter out expired and parked domains
|
|
if any(filtered_substr in str(ns) for filtered_substr in NS_FILTER_SUBSTRINGS):
|
|
return domain
|
|
|
|
except (dns.exception.Timeout, dns.resolver.NXDOMAIN,
|
|
dns.resolver.YXDOMAIN, dns.resolver.NoNameservers):
|
|
return domain
|
|
except dns.resolver.NoAnswer:
|
|
# Do not thread domain as broken if the answer is empty,
|
|
# but do if it's cloudfront subdomain
|
|
if domain.endswith('.cloudfront.net'):
|
|
return domain
|
|
|
|
|
|
def tasksProvider(domainiter, resolver):
|
|
for domain in domainiter:
|
|
yield asyncio.ensure_future(resolver.nxresolve(domain.strip()))
|
|
|
|
async def runTasksWithProgress(tasks, tasknumber, concurrent_tasks):
|
|
progress = 0
|
|
old_progress = 0
|
|
ret = []
|
|
|
|
# this line actually "runs" coroutine
|
|
tasklist = list(itertools.islice(tasks, concurrent_tasks))
|
|
tasklist_next = []
|
|
while tasklist:
|
|
for task in asyncio.as_completed(tasklist):
|
|
ret.append(await task)
|
|
progress = int(len(ret) / tasknumber * 100)
|
|
if old_progress < progress:
|
|
print("{}%...".format(progress), end='\r', file=sys.stderr, flush=True)
|
|
old_progress = progress
|
|
|
|
for newtask in tasks: # this line actually "runs" coroutine
|
|
tasklist_next.append(newtask)
|
|
break
|
|
tasklist = tasklist_next
|
|
tasklist_next = []
|
|
|
|
print(file=sys.stderr)
|
|
return ret
|
|
|
|
async def main():
|
|
if len(sys.argv) != 2:
|
|
print("Incorrect arguments!")
|
|
sys.exit(1)
|
|
|
|
r = AZResolver()
|
|
r.limitConcurrency(INITIAL_PASS_CONCURRENCY)
|
|
r.timeout = INITIAL_PASS_TIMEOUT
|
|
r.lifetime = INITIAL_PASS_TIMEOUT
|
|
|
|
# Load domain file list and schedule resolving
|
|
domaincount = 0
|
|
try:
|
|
with open(sys.argv[1], 'r') as domainlist:
|
|
for domain in domainlist:
|
|
domaincount += 1
|
|
except OSError as e:
|
|
print("Can't open file", sys.argv[1], e, file=sys.stderr)
|
|
sys.exit(2)
|
|
|
|
print("Loaded list of {} elements, resolving NXDOMAINS".format(domaincount), file=sys.stderr)
|
|
#sys.exit(0)
|
|
|
|
try:
|
|
domainfile = open(sys.argv[1], 'r')
|
|
|
|
# Resolve domains, first try
|
|
nxresolved_first = await runTasksWithProgress(
|
|
tasksProvider(domainfile, r), domaincount,
|
|
INITIAL_PASS_CONCURRENCY * TASK_AMOUNT_MULTIPLIER
|
|
)
|
|
nxresolved_first = list(filter(None, nxresolved_first))
|
|
|
|
print("Got {} broken domains, trying to resolve them again "
|
|
"to make sure".format(len(nxresolved_first)), file=sys.stderr)
|
|
|
|
# Second try
|
|
r.limitConcurrency(FINAL_PASS_CONCURRENCY)
|
|
r.timeout = FINAL_PASS_TIMEOUT
|
|
r.lifetime = FINAL_PASS_TIMEOUT
|
|
|
|
nxresolved_second = await runTasksWithProgress(
|
|
tasksProvider(nxresolved_first, r), len(nxresolved_first),
|
|
FINAL_PASS_CONCURRENCY * TASK_AMOUNT_MULTIPLIER
|
|
)
|
|
nxresolved_second = list(filter(None, nxresolved_second))
|
|
|
|
print("Finally, got {} broken domains".format(len(nxresolved_second)), file=sys.stderr)
|
|
for domain in nxresolved_second:
|
|
print(domain)
|
|
|
|
domainfile.close()
|
|
|
|
except (SystemExit, KeyboardInterrupt):
|
|
print("Exiting")
|
|
|
|
|
|
if __name__ == '__main__':
|
|
if dns.__version__ == '2.0.0':
|
|
# Monkey-patch dnspython 2.0.0 bug #572
|
|
# https://github.com/rthalley/dnspython/issues/572
|
|
class monkeypatched_DatagramProtocol(dns._asyncio_backend._DatagramProtocol):
|
|
def error_received(self, exc): # pragma: no cover
|
|
if self.recvfrom and not self.recvfrom.done():
|
|
self.recvfrom.set_exception(exc)
|
|
|
|
def connection_lost(self, exc):
|
|
if self.recvfrom and not self.recvfrom.done():
|
|
self.recvfrom.set_exception(exc)
|
|
|
|
dns._asyncio_backend._DatagramProtocol = monkeypatched_DatagramProtocol
|
|
|
|
try:
|
|
asyncio.run(main())
|
|
except (SystemExit, KeyboardInterrupt):
|
|
sys.exit(3)
|