File size: 2,579 Bytes
fb08e0a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 |
#!/usr/bin/env python3
# License: MIT
# Copyright (C) 2024, Shinon.
# Retrieves a full list of wikis to be scraped
import asyncio
import csv
import urllib.parse
from bs4 import BeautifulSoup
from proxy_magic_session import get_async_session
async def main():
with open("fandom_wikis_210224.csv", "w", newline="") as f:
writer = csv.writer(f)
session = get_async_session()
root = "https://community.fandom.com"
r = await session.get(f"{root}/wiki/Special:NewWikis?limit=500")
if r.status_code == 200:
nrow = 0
soup = BeautifulSoup(r.text, "lxml")
for doms in soup.select(".mw-spcontent li > a"):
href: str = doms.get("href", "")
if href:
parsed = urllib.parse.urlparse(href)
domain = parsed.netloc
if parsed.path.strip("/"):
sp = f"{parsed.path}wiki/"
else:
sp = "/wiki/"
print("Add wiki:", domain, "|", doms.get_text().strip())
writer.writerow([domain, doms.get_text().strip(), sp, 0])
nrow += 1
next_page = soup.find("a", attrs={"rel": "next", "class": "mw-nextlink"})
if next_page:
next_page_url = f'{root}{next_page.get("href")}'
else:
next_page_url = None
while next_page_url:
nrow = 0
r = await session.get(next_page_url)
soup = BeautifulSoup(r.text, "lxml")
for doms in soup.select(".mw-spcontent li > a"):
href: str = doms.get("href", "")
if href:
parsed = urllib.parse.urlparse(href)
domain = parsed.netloc
if parsed.path.strip("/"):
sp = f"{parsed.path}wiki/"
else:
sp = "/wiki/"
print("Add wiki:", domain, "|", doms.get_text().strip())
writer.writerow([domain, doms.get_text().strip(), sp, 0])
nrow += 1
next_page = soup.find(
"a", attrs={"rel": "next", "class": "mw-nextlink"}
)
if next_page:
next_page_url = f'{root}{next_page.get("href")}'
else:
next_page_url = None
print(next_page_url)
asyncio.run(main())
|