# lld/crawler.py import asyncio from urllib.parse import urlparse import aiohttp from extraction import Extractor from tqdm.asyncio import tqdm, tqdm_asyncio def get_domain(url: str): parsed = urlparse(url) return parsed.netloc async def gather_with_concurrency(n: int, tasks): semaphore = asyncio.Semaphore(n) async def sem_task(task): async with semaphore: return await task results = asyncio.as_completed(tasks) return results async def scrape_description( url: str, session: aiohttp.ClientSession, timeout: int = 15 ) -> str: """Scrape description of the given url from and tags.""" extractor = Extractor() domain = get_domain(url) try: async with session.get(url, timeout=timeout) as response: html = await response.text() data = extractor.extract(html, source_url=url) collected = [ *data.titles, data.description, domain, ] results = "\n".join(filter(bool, collected)) return results except: return domain async def crawl(urls: list[str], batch_size: int = 100): connector = aiohttp.TCPConnector(limit=None, ttl_dns_cache=300) async with aiohttp.ClientSession(connector=connector) as session: gen_results = await gather_with_concurrency( batch_size, [scrape_description(url, session) for url in urls] ) for result in gen_results: yield await result async def run(urls: list[str], batch_size: int = 100): crawler = crawl(urls, batch_size) results = [c async for c in tqdm_asyncio(crawler, total=len(urls))] return results