|
import os |
|
import re |
|
import requests |
|
import pandas as pd |
|
from bs4 import BeautifulSoup |
|
|
|
url = "https://samples.vx-underground.org/samples/Families/" |
|
html_content = requests.get(url).text |
|
soup = BeautifulSoup(html_content, "html.parser") |
|
|
|
family_links = [a["href"] for a in soup.find_all("a", href=True) if re.match(r'https://samples.vx-underground.org/samples/Families/[\w-]+/', a["href"])] |
|
|
|
data = [] |
|
download_path = os.getcwd() |
|
|
|
for family_link in family_links: |
|
family_name = family_link.split('/')[-2] |
|
print(f'Downloading files for {family_name}...') |
|
family_html_content = requests.get(family_link).text |
|
family_soup = BeautifulSoup(family_html_content, "html.parser") |
|
|
|
sample_links = [a["href"] for a in family_soup.find_all("a", href=True) if a["href"].endswith(".7z")] |
|
|
|
for sample_link in sample_links: |
|
sample_name = sample_link.split('/')[-1] |
|
|
|
family_path = os.path.join(download_path, family_name) |
|
if not os.path.exists(family_path): |
|
os.makedirs(family_path) |
|
file_path = os.path.join(family_path, sample_name) |
|
|
|
if not os.path.exists(file_path): |
|
try: |
|
response = requests.get(sample_link, stream=True) |
|
with open(file_path, "wb") as f: |
|
f.write(response.content) |
|
except Exception as e: |
|
print(f"Error downloading {sample_name}: {e}") |
|
continue |
|
|
|
data.append({"family": family_name, "sample": sample_name}) |
|
|
|
df = pd.DataFrame(data) |
|
df.to_csv("samples_and_families.csv", index=False) |
|
|