|
|
|
from transformers import pipeline |
|
from bs4 import BeautifulSoup |
|
import requests |
|
import urllib |
|
import json |
|
import argparse |
|
from pathlib import Path |
|
import os |
|
|
|
from controlnet_aux.processor import Processor |
|
from colorthief import ColorThief |
|
from colors import name_that_color, rgb_to_hex, NTC_NAMES, CSS3_NAMES |
|
|
|
from PIL import Image |
|
|
|
|
|
class Scraper(object): |
|
def __init__(self, urls): |
|
self.urls = urls |
|
|
|
def fetch(self, url): |
|
return requests.get(url, headers={"User-Agent": "Mozilla/5.0"}) |
|
|
|
def parse(self, text): |
|
return BeautifulSoup(text, 'html.parser') |
|
|
|
def download_image(self, url, path): |
|
image = requests.get(url) |
|
open(path, 'wb').write(image.content) |
|
|
|
|
|
class GraffitiDatabaseScraper(Scraper): |
|
def __init__(self): |
|
Scraper.__init__(self, ["https://graffiti-database.com"]) |
|
|
|
def get_image_urls(self, parsed): |
|
pages = parsed.select('.image-info a') |
|
image_urls = [] |
|
for page in pages: |
|
image_urls.append(page.get('href')) |
|
return image_urls |
|
|
|
def scrape_image(self, url): |
|
print("Scraping image page {0}".format(url)) |
|
|
|
fetched = self.fetch(url) |
|
|
|
if fetched.status_code != 200: |
|
raise "Scraping image page did fail" |
|
|
|
text = fetched.text |
|
parsed = self.parse(text) |
|
|
|
image = parsed.select('.img-fluid')[0].get('src') |
|
|
|
tag_links = parsed.select('a.tag') |
|
|
|
url_parts = urllib.parse.unquote( |
|
urllib.parse.unquote(url)).rsplit("/", 2) |
|
artist = url_parts[1] |
|
return { |
|
"image_url": image, |
|
"city": tag_links[-1].get_text().strip(), |
|
"artist": artist if artist != "Writer Unknown" and artist != "Writers Unknown" else None |
|
} |
|
|
|
def scrape_page(self, url): |
|
print("Scraping page {0}".format(url)) |
|
|
|
text = self.fetch(url).text |
|
parsed = self.parse(text) |
|
|
|
image_urls = self.get_image_urls(parsed) |
|
|
|
if len(image_urls) == 0: |
|
return False |
|
|
|
for image_url in image_urls: |
|
try: |
|
result = self.scrape_image( |
|
"https://graffiti-database.com" + image_url) |
|
except: |
|
continue |
|
|
|
file_name = result["image_url"].split('/')[-1] |
|
|
|
self.download_image(result["image_url"], |
|
'./images/' + file_name) |
|
|
|
result["file"] = file_name |
|
|
|
with open('./images/' + file_name + '.json', 'w', encoding='utf-8') as f: |
|
f.write(json.dumps(result, indent=2, ensure_ascii=False)) |
|
|
|
return True |
|
|
|
def scrape(self): |
|
url = self.urls[0] |
|
print("Scraping {0}".format( |
|
url)) |
|
|
|
count = 1 |
|
has_images = True |
|
while has_images: |
|
has_images = self.scrape_page(url + "?page=" + str(count)) |
|
count += 1 |
|
|
|
|
|
class GraffitiScraper(Scraper): |
|
def __init__(self): |
|
Scraper.__init__(self, ["https://www.graffiti.org/index/world.html", |
|
"https://www.graffiti.org/index/europe.html", "https://www.graffiti.org/index/usa.html"]) |
|
|
|
def scrape_page(self, url, city_name): |
|
print("Scraping page {0}".format(url)) |
|
|
|
text = self.fetch(url).text |
|
parsed = self.parse(text) |
|
|
|
image_elements = parsed.select('a[href*=".jpg"]') |
|
|
|
for image_element in image_elements: |
|
if image_element.get_text().strip() == "": |
|
tags = image_element.find_next_sibling(string=True).get_text( |
|
).strip() if image_element.find_next_sibling(string=True) else "" |
|
else: |
|
continue |
|
image_url = image_element.get("href").replace("/", "_") |
|
url_parts = url.split('/') |
|
url_parts[-1] = image_element.get("href") |
|
self.download_image("/".join(url_parts), |
|
'./images/' + image_url) |
|
|
|
with open('./images/' + image_url + '.json', 'w', encoding='utf-8') as f: |
|
f.write(json.dumps({ |
|
"file": image_url, |
|
"image_url": "/".join(url_parts), |
|
"artist": tags, |
|
"city": city_name if city_name != "Various cities" else None |
|
}, indent=2, ensure_ascii=False)) |
|
|
|
def scrape_url(self, url): |
|
print("Scraping url {0}".format(url)) |
|
|
|
text = self.fetch(url).text |
|
parsed = self.parse(text) |
|
|
|
cities = parsed.find_all("h4") |
|
|
|
for city in cities: |
|
city_name = city.get_text().split("\n")[0].strip() |
|
pages = city.find_all("a") |
|
for page in pages: |
|
if page.get_text().strip() == "§": |
|
continue |
|
self.scrape_page( |
|
"https://www.graffiti.org/index/" + page.get("href"), city_name) |
|
|
|
def scrape(self): |
|
for url in self.urls: |
|
self.scrape_url(url) |
|
|
|
|
|
class CLI(): |
|
def __init__(self): |
|
parser = argparse.ArgumentParser( |
|
prog='graffiti-cli', |
|
description='Tools for setting up the dataset') |
|
subparsers = parser.add_subparsers(dest="command", required=True) |
|
|
|
scrape = subparsers.add_parser( |
|
'scrape', help='Scrapes data sources and downloads images') |
|
scrape.add_argument('--source', |
|
default='graffiti.org', |
|
choices=['graffiti.org', 'graffiti-database.com'], |
|
help='Choose data source to scrape') |
|
subparsers.add_parser('cleanup', help='Cleans up downloaded images') |
|
subparsers.add_parser('caption', help='Captions downloaded images') |
|
subparsers.add_parser('palette', help='Creates color palettes for downloaded images') |
|
subparsers.add_parser('condition', help='Condition downloaded images') |
|
metadata = subparsers.add_parser('metadata', help='Creates single meta files from metadata.jsonl') |
|
metadata.add_argument('--source', |
|
default='graffiti.org', |
|
choices=['graffiti.org', 'graffiti-database.com'], |
|
help='Choose data source to use') |
|
|
|
args = parser.parse_args() |
|
if args.command == 'scrape': |
|
if args.source == 'graffiti.org': |
|
GraffitiScraper().scrape() |
|
elif args.source == 'graffiti-database.com': |
|
GraffitiDatabaseScraper().scrape() |
|
elif args.command == 'cleanup': |
|
path = Path("./images").rglob("*.jpg") |
|
for i, img_p in enumerate(path): |
|
try: |
|
Image.open(img_p).load() |
|
except Exception: |
|
path_name = str(img_p) |
|
print(path_name + " is broken. Deleting!") |
|
os.remove(path_name) |
|
os.remove(path_name + ".json") |
|
elif args.command == 'caption': |
|
captioner = pipeline( |
|
"image-to-text", model="Salesforce/blip-image-captioning-base") |
|
path = Path("./images").rglob("*.jpg") |
|
count = len(list(Path("./images").rglob("*.jpg"))) |
|
for i, img_p in enumerate(path): |
|
path_name = str(img_p) |
|
with open(path_name + ".json", 'r+', encoding='utf-8') as f: |
|
data = json.load(f) |
|
|
|
f.seek(0) |
|
|
|
caption = captioner(path_name)[0]["generated_text"] |
|
|
|
if "album" in caption: |
|
caption = "a wall with graffiti on it" |
|
|
|
json.dump({ |
|
"file": data["file"], |
|
"image_url": data["image_url"], |
|
"artist": data["artist"], |
|
"city": data["city"], |
|
"caption": caption |
|
}, f, indent=2) |
|
f.truncate() |
|
|
|
print("{0} / {1}".format(i + 1, count), path_name, caption) |
|
elif args.command == 'palette': |
|
path = Path("./images").rglob("*.jpg") |
|
count = len(list(Path("./images").rglob("*.jpg"))) |
|
for i, img_p in enumerate(path): |
|
path_name = str(img_p) |
|
with open(path_name + ".json", 'r+', encoding='utf-8') as f: |
|
data = json.load(f) |
|
|
|
f.seek(0) |
|
|
|
color_thief = ColorThief(path_name) |
|
|
|
palette = [] |
|
for color in color_thief.get_palette(color_count=6): |
|
color_hex = rgb_to_hex(color) |
|
palette.append([color_hex, name_that_color(color_hex, NTC_NAMES), name_that_color(color_hex, CSS3_NAMES)]) |
|
|
|
|
|
json.dump({ |
|
"file": data["file"], |
|
"image_url": data["image_url"], |
|
"artist": data["artist"], |
|
"city": data["city"], |
|
"caption": data["caption"], |
|
"palette": palette |
|
}, f, indent=2) |
|
f.truncate() |
|
|
|
print("{0} / {1}".format(i + 1, count), path_name, palette) |
|
elif args.command == 'metadata': |
|
with open("data/" + args.source + "/metadata.jsonl", encoding="utf-8") as f: |
|
for row in f: |
|
data = json.loads(row) |
|
with open('./images/' + data["file"] + '.json', 'w') as j: |
|
j.write(json.dumps(data, indent=2, ensure_ascii=False)) |
|
elif args.command == 'conditioning': |
|
processor_id = 'softedge_hed' |
|
processor = Processor(processor_id) |
|
|
|
path = Path("./images").rglob("*.jpg") |
|
count = len(list(Path("./images").rglob("*.jpg"))) |
|
for i, img_p in enumerate(path): |
|
path_name = str(img_p) |
|
output_path = path_name.replace("images/", "conditioning/") |
|
img = Image.open(img_p).convert("RGB") |
|
|
|
processed_image = processor(img, to_pil=True) |
|
processed_image.save(output_path) |
|
|
|
print("{0} / {1}".format(i + 1, count), path_name, output_path) |
|
|
|
def main(): |
|
CLI() |
|
|
|
|
|
if __name__ == "__main__": |
|
main() |
|
|