from transformers import pipeline from bs4 import BeautifulSoup import requests import urllib import json import argparse from pathlib import Path import os from controlnet_aux.processor import Processor from colorthief import ColorThief from colors import name_that_color, rgb_to_hex, NTC_NAMES, CSS3_NAMES from PIL import Image class Scraper(object): def __init__(self, urls): self.urls = urls def fetch(self, url): return requests.get(url, headers={"User-Agent": "Mozilla/5.0"}) def parse(self, text): return BeautifulSoup(text, 'html.parser') def download_image(self, url, path): image = requests.get(url) open(path, 'wb').write(image.content) class GraffitiDatabaseScraper(Scraper): def __init__(self): Scraper.__init__(self, ["https://graffiti-database.com"]) def get_image_urls(self, parsed): pages = parsed.select('.image-info a') image_urls = [] for page in pages: image_urls.append(page.get('href')) return image_urls def scrape_image(self, url): print("Scraping image page {0}".format(url)) fetched = self.fetch(url) if fetched.status_code != 200: raise "Scraping image page did fail" text = fetched.text parsed = self.parse(text) image = parsed.select('.img-fluid')[0].get('src') tag_links = parsed.select('a.tag') url_parts = urllib.parse.unquote( urllib.parse.unquote(url)).rsplit("/", 2) artist = url_parts[1] return { "image_url": image, "city": tag_links[-1].get_text().strip(), "artist": artist if artist != "Writer Unknown" and artist != "Writers Unknown" else None } def scrape_page(self, url): print("Scraping page {0}".format(url)) text = self.fetch(url).text parsed = self.parse(text) image_urls = self.get_image_urls(parsed) if len(image_urls) == 0: return False for image_url in image_urls: try: result = self.scrape_image( "https://graffiti-database.com" + image_url) except: continue file_name = result["image_url"].split('/')[-1] self.download_image(result["image_url"], './images/' + file_name) result["file"] = file_name with open('./images/' + file_name + '.json', 'w', encoding='utf-8') as f: f.write(json.dumps(result, indent=2, ensure_ascii=False)) return True def scrape(self): url = self.urls[0] print("Scraping {0}".format( url)) count = 1 has_images = True while has_images: has_images = self.scrape_page(url + "?page=" + str(count)) count += 1 class GraffitiScraper(Scraper): def __init__(self): Scraper.__init__(self, ["https://www.graffiti.org/index/world.html", "https://www.graffiti.org/index/europe.html", "https://www.graffiti.org/index/usa.html"]) def scrape_page(self, url, city_name): print("Scraping page {0}".format(url)) text = self.fetch(url).text parsed = self.parse(text) image_elements = parsed.select('a[href*=".jpg"]') for image_element in image_elements: if image_element.get_text().strip() == "": tags = image_element.find_next_sibling(string=True).get_text( ).strip() if image_element.find_next_sibling(string=True) else "" else: continue image_url = image_element.get("href").replace("/", "_") url_parts = url.split('/') url_parts[-1] = image_element.get("href") self.download_image("/".join(url_parts), './images/' + image_url) with open('./images/' + image_url + '.json', 'w', encoding='utf-8') as f: f.write(json.dumps({ "file": image_url, "image_url": "/".join(url_parts), "artist": tags, "city": city_name if city_name != "Various cities" else None }, indent=2, ensure_ascii=False)) def scrape_url(self, url): print("Scraping url {0}".format(url)) text = self.fetch(url).text parsed = self.parse(text) cities = parsed.find_all("h4") for city in cities: city_name = city.get_text().split("\n")[0].strip() pages = city.find_all("a") for page in pages: if page.get_text().strip() == "ยง": continue self.scrape_page( "https://www.graffiti.org/index/" + page.get("href"), city_name) def scrape(self): for url in self.urls: self.scrape_url(url) class CLI(): def __init__(self): parser = argparse.ArgumentParser( prog='graffiti-cli', description='Tools for setting up the dataset') subparsers = parser.add_subparsers(dest="command", required=True) scrape = subparsers.add_parser( 'scrape', help='Scrapes data sources and downloads images') scrape.add_argument('--source', default='graffiti.org', choices=['graffiti.org', 'graffiti-database.com'], help='Choose data source to scrape') subparsers.add_parser('cleanup', help='Cleans up downloaded images') subparsers.add_parser('caption', help='Captions downloaded images') subparsers.add_parser('palette', help='Creates color palettes for downloaded images') subparsers.add_parser('condition', help='Condition downloaded images') metadata = subparsers.add_parser('metadata', help='Creates single meta files from metadata.jsonl') metadata.add_argument('--source', default='graffiti.org', choices=['graffiti.org', 'graffiti-database.com'], help='Choose data source to use') args = parser.parse_args() if args.command == 'scrape': if args.source == 'graffiti.org': GraffitiScraper().scrape() elif args.source == 'graffiti-database.com': GraffitiDatabaseScraper().scrape() elif args.command == 'cleanup': path = Path("./images").rglob("*.jpg") for i, img_p in enumerate(path): try: Image.open(img_p).load() except Exception: path_name = str(img_p) print(path_name + " is broken. Deleting!") os.remove(path_name) os.remove(path_name + ".json") elif args.command == 'caption': captioner = pipeline( "image-to-text", model="Salesforce/blip-image-captioning-base") path = Path("./images").rglob("*.jpg") count = len(list(Path("./images").rglob("*.jpg"))) for i, img_p in enumerate(path): path_name = str(img_p) with open(path_name + ".json", 'r+', encoding='utf-8') as f: data = json.load(f) f.seek(0) caption = captioner(path_name)[0]["generated_text"] if "album" in caption: caption = "a wall with graffiti on it" json.dump({ "file": data["file"], "image_url": data["image_url"], "artist": data["artist"], "city": data["city"], "caption": caption }, f, indent=2) f.truncate() print("{0} / {1}".format(i + 1, count), path_name, caption) elif args.command == 'palette': path = Path("./images").rglob("*.jpg") count = len(list(Path("./images").rglob("*.jpg"))) for i, img_p in enumerate(path): path_name = str(img_p) with open(path_name + ".json", 'r+', encoding='utf-8') as f: data = json.load(f) f.seek(0) color_thief = ColorThief(path_name) palette = [] for color in color_thief.get_palette(color_count=6): color_hex = rgb_to_hex(color) palette.append([color_hex, name_that_color(color_hex, NTC_NAMES), name_that_color(color_hex, CSS3_NAMES)]) json.dump({ "file": data["file"], "image_url": data["image_url"], "artist": data["artist"], "city": data["city"], "caption": data["caption"], "palette": palette }, f, indent=2) f.truncate() print("{0} / {1}".format(i + 1, count), path_name, palette) elif args.command == 'metadata': with open("data/" + args.source + "/metadata.jsonl", encoding="utf-8") as f: for row in f: data = json.loads(row) with open('./images/' + data["file"] + '.json', 'w') as j: j.write(json.dumps(data, indent=2, ensure_ascii=False)) elif args.command == 'conditioning': processor_id = 'softedge_hed' processor = Processor(processor_id) path = Path("./images").rglob("*.jpg") count = len(list(Path("./images").rglob("*.jpg"))) for i, img_p in enumerate(path): path_name = str(img_p) output_path = path_name.replace("images/", "conditioning/") img = Image.open(img_p).convert("RGB") processed_image = processor(img, to_pil=True) processed_image.save(output_path) print("{0} / {1}".format(i + 1, count), path_name, output_path) def main(): CLI() if __name__ == "__main__": main()