# Created by Danyang Zhang @X-Lance from mitmproxy import http from mitmproxy import ctx import os.path import functools import classify_url import load_response import datetime import locale import gzip import lxml.html import lxml.cssselect from pyserini.search.lucene import LuceneSearcher import csv import urllib.parse from typing import List from typing import Mapping from mitmproxy.coretypes.multidict import MultiDict import random class Replayer: # class `Replayer` {{{ # def __init__(self, start_cache_index: int, replay_path: str, template_path: str, index_path: str, meta_path: str): self.cache_index: int = start_cache_index self.replay_path: str = replay_path self.template_path: str = template_path self.index_path: str = index_path self.meta_path: str = meta_path with open(meta_path) as f: reader = csv.DictReader(f) self.meta_database: Dict[str, Mapping[str, str]]\ = {itm["doc_id"]: itm for itm in reader} self.doc_list: List[str] = list(self.meta_database.keys()) locale.setlocale(locale.LC_TIME, "en_US.utf8") self.dateformat: str = "%a, %d %b %Y %H:%M:%S GMT" self.searcher: LuceneSearcher = LuceneSearcher(self.index_path) # CSS Selectors {{{ # self.search_input_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("input#hs_query", translator="html") self.result_list_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("div#searchresults_list.wh_block", translator="html") self.result_list_anchor_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("div#search_adblock_bottom", translator="html") self.result_footer_anchor_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("div#searchresults_footer>div.sr_text", translator="html") self.result_thumb_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("div.result_thumb", translator="html") self.result_title_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("div.result_title", translator="html") self.result_view_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("li.sr_view", translator="html") self.result_updated_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("li.sr_updated", translator="html") self.result_verif_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("li.sp_verif", translator="html") self.sha_index_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("input[name=\"sha_index\"]", translator="html") self.sha_id_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("input[name=\"sha_id\"]", translator="html") self.sha_title_selector: lxml.cssselect.CSSSelector\ = lxml.cssselect.CSSSelector("input[name=\"sha_title\"]", translator="html") # }}} CSS Selectors # self.search_page_capacity: int = 30 self.search_capacity: int = 300 def request(self, flow: http.HTTPFlow): if flow.request.pretty_host=="www.wikihow.com": url_path = flow.request.path url_key = classify_url.classify(url_path) if url_key in [ r"/x/collect\?t={first,later}&*" , r"/x/collect\?t={exit,amp}&*" , r"/x/amp-view\?*" , r"/ev/*" ]: # Control Flows {{{ # headers = {} response_time = datetime.datetime.utcnow() headers["retry-after"] = "0" headers["accept-ranges"] = "bytes" headers["date"] = response_time.strftime(self.dateformat) headers["x-timer"] = "S{:.6f},VS0,VE0".format(response_time.timestamp()) headers["x-c"] = "cache-tyo{:d}-TYO,M".format(self.cache_index) headers["x-content-type-options"] = "nosniff" headers["x-xss-protection"] = "1; mode=block" headers["strict-transport-security"] = "max-age=31536000; includeSubDomains; preload" headers["content-type"] = "text/plain; charset=UTF-8" headers["x-robots-tag"] = "noindex, nofollow" flow.response = http.Response.make(204, headers=headers) self.cache_index += 1 # }}} Control Flows # elif url_path.startswith("/wikiHowTo?search="): # Search Pages {{{ # url_items = urllib.parse.urlparse(url_path) queries = urllib.parse.parse_qs(url_items.query) search_keywords = queries["search"][0] start_index = int(queries.get("start", ["0"])[0]) hits = self.searcher.search(search_keywords, k=self.search_capacity) response_time = datetime.datetime.utcnow() # build the webpage page_body = lxml.html.parse( os.path.join(self.template_path, "search-page.html.template")).getroot() # 1. fill the parameters in the page body search_input = self.search_input_selector(page_body)[0] search_input.set("value", search_keywords) # 2. prepare result list #result_list = self.result_list_selector(page_body)[0] result_list_bottom = self.result_list_anchor_selector(page_body)[0] with open(os.path.join(self.template_path, "search-item.html.template")) as f: result_item_html = "".join(f.readlines()) for i, h in zip(range(start_index, start_index+self.search_page_capacity), hits[start_index:start_index+self.search_page_capacity]): docid = h.docid article_path = docid.replace("%2f", "/") result_item = lxml.html.fromstring(result_item_html) # Update Item Parameters {{{ # result_item.set("href", "https://www.wikihow.com/{:}".format(article_path)) result_thumb = self.result_thumb_selector(result_item)[0] thumb_url = self.meta_database[docid]["thumb_url"] if thumb_url!="": result_thumb.set("style", "background-image: url(https://www.wikihow.com{:})".format(thumb_url.replace("%2f", "/"))) else: result_thumb.set("style", "background-image: url(https://www.wikihow.com/5/5f/{doc_id:}-Step-2.jpg/-crop-250-145-193px-{doc_id}-Step-2.jpg)".format(doc_id=article_path)) new_result_title = lxml.html.fromstring( "
{:}
"\ .format(" ".join( map( lambda w: "" + w + "" , self.meta_database[docid]["title"].split() ) ) ) ) result_title = self.result_title_selector(result_item)[0] result_title.getparent().replace(result_title, new_result_title) result_view = self.result_view_selector(result_item)[0] view_counts = self.meta_database[docid]["sr_view"] if view_counts!="": result_view.text = "{:} views\t\t\t\t\t\t".format(view_counts) else: result_view.text = "0 views\t\t\t\t\t\t" result_updated = self.result_updated_selector(result_item)[0] updated_date = self.meta_database[docid]["sr_updated"] if updated_date!="": updated_date = datetime.datetime.strptime(updated_date, "%B %d, %Y") updating_duration = response_time - updated_date # Calculate Time Diff String {{{ # days = updating_duration.days if days<7: time_diff_str = "{:} day{:} ago".format( days, "" if days==1 else "s") elif days<30: weeks = days // 7 time_diff_str = "{:} week{:} ago".format( weeks, "" if weeks==1 else "s") elif days<365: months = days // 30 time_diff_str = "{:} month{:} ago".format( months, "" if months==1 else "s") else: years = days // 365 time_diff_str = "{:} year{:} ago".format( years, "" if years==1 else "s") # }}} Calculate Time Diff String # list(result_updated)[0].tail = time_diff_str + "\t\t\t\t\t\t" else: list(result_updated)[0].tail = "12 hours ago\t\t\t\t\t\t" result_verif = self.result_verif_selector(result_item)[-1] verif_type = self.meta_database[docid]["sr_verif"] if verif_type=="E": result_verif.text = "Expert Co-Authored\t\t\t\t\t\t\t" elif verif_type=="Q": result_verif.text = "Quality Tested\t\t\t\t\t\t\t" else: result_verif.getparent().remove(result_verif) sha_index = self.sha_index_selector(result_item)[0] sha_index.set("value", str(i+1)) sha_id = self.sha_id_selector(result_item)[0] sha_id.set("value", self.meta_database[docid]["sha_id"]) sha_title = self.sha_title_selector(result_item)[0] sha_title.set("value", article_path) # }}} Update Item Parameters # result_list_bottom.addprevious(result_item) # 3. prepare footer result_footer_bottom = self.result_footer_anchor_selector(page_body)[0] with open(os.path.join(self.template_path, "search-button.html.template")) as f: result_footer = lxml.html.fromstring("".join(f.readlines())) next_button, previous_button, statistics_text = list(result_footer) if start_index+self.search_page_capacityNext >") result_footer_bottom.addprevious(disabled_next) if start_index>0: # has previous enabled_previous = lxml.html.fromstring( "< Previous") enabled_previous.set("href", "/wikiHowTo?search={:}&start={:}&wh_an=1".format( urllib.parse.quote_plus(search_keywords), start_index-self.search_page_capacity)) result_footer_bottom.addprevious(enabled_previous) else: result_footer_bottom.addprevious(previous_button) statistics_text.text = "{:} Results".format(self.search_capacity) result_footer_bottom.addprevious(statistics_text) # 4. return result full_page = lxml.html.tostring(page_body) headers = {} headers["content-type"] = "text/html; charset=UTF-8" headers["content-language"] = "en" headers["x-frame-options"] = "SAMEORIGIN" headers["x-p"] = "ma" headers["expires"] = (response_time + datetime.timedelta(days=1))\ .strftime(self.dateformat) headers["cache-control"] = "s-maxage=86400, must-revalidate, max-age=0" headers["content-encoding"] = "gzip" headers["accept-ranges"] = "bytes" headers["date"] = response_time.strftime(self.dateformat) headers["age"] = "0" headers["x-timer"] = "S{:.6f},VS0,VE0".format(response_time.timestamp()) headers["x-c"] = "cache-tyo{:d}-TYO,M".format(self.cache_index) headers["x-content-type-options"] = "nosniff" headers["x-xss-protection"] = "1; mode=block" headers["strict-transport-security"] = "max-age=31536000; includeSubDomains; preload" headers["set-cookie"] =\ "whv=lbYVTnTp1cUHHoDGDOwR; expires={:}; domain=.wikihow.com; path=/; secure"\ .format((response_time + datetime.timedelta(days=3654)).strftime(self.dateformat)) headers["vary"] = "Cookie, Accept-Encoding" flow.response = http.Response.make(200, content=full_page, headers=headers) self.cache_index += 1 # }}} Search Pages # elif url_path.startswith("/Special:Randomizer"): # Random Page {{{ # random_index = random.randrange(len(self.doc_list)) doc_id = self.doc_list[random_index] article_path = doc_id.replace("%2f", "/") headers = {} response_time = datetime.datetime.utcnow() headers["content-type"] = "text/html; charset=UTF-8" headers["expires"] = "Thu, 01 Jan 1970 00:00:00 GMT" headers["cache-control"] = "private, must-revalidate, max-age=0" headers["x-p"] = "ma" headers["location"] = "https://www.wikihow.com/{:}".format(article_path) headers["content-encoding"] = "gzip" headers["accept-ranges"] = "bytes" headers["date"] = response_time.strftime(self.dateformat) headers["x-timer"] = "S{:.6f},VS0,VE0".format(response_time.timestamp()) headers["x-c"] = "cache-tyo{:d}-TYO,M".format(self.cache_index) headers["x-content-type-options"] = "nosniff" headers["x-xss-protection"] = "1; mode=block" headers["strict-transport-security"] = "max-age=31536000; includeSubDomains; preload" headers["vary"] = "Cookie, Accept-Encoding" flow.response = http.Response.make(302, headers=headers) self.cache_index += 1 # }}} Random Page # elif url_path.startswith("/Special:RateItem"): # Reting Flows {{{ # headers = {} response_time = datetime.datetime.utcnow() headers["content-type"] = "text/html; charset=UTF-8" headers["access-control-allow-credentials"] = "true" headers["access-control-allow-origin"] = "https://www.wikihow.com" headers["access-control-expose-headers"] = "AMP-Access-Control-Allow-Source-Origin" headers["amp-access-control-allow-source-origin"] = "https://www.wikihow.com" #headers["set-cookie"] =\ #"UseDC=master; expires={:}; Max-Age=10; path=/; domain=www.wikihow.com; secure; HttpOnly" headers["content-language"] = "en" headers["x-frame-options"] = "SAMEORIGIN" headers["expires"] = "Thu, 01 Jan 1970 00:00:00 GMT" headers["cache-control"] = "private, must-revalidate, max-age=0" headers["x-p"] = "ck ma" headers["content-encoding"] = "gzip" headers["accept-ranges"] = "bytes" headers["date"] = response_time.strftime(self.dateformat) headers["x-timer"] = "S{:.6f},VS0,VE0".format(response_time.timestamp()) headers["x-c"] = "cache-tyo{:d}-TYO,M".format(self.cache_index) headers["x-content-type-options"] = "nosniff" headers["x-xss-protection"] = "1; mode=block" headers["strict-transport-security"] = "max-age=31536000; includeSubDomains; preload" headers["vary"] = "Cookie, Accept-Encoding" flow.response = http.Response.make(200, content=b'{"result":"true"}', headers=headers) cookies_attribute = MultiDict([ ("expires", (response_time + datetime.timedelta(seconds=10))\ .strftime(self.dateformat)), ("Max-Age", "10"), ("path", "/"), ("domain", "www.wikihow.com"), ("secure", None), ("HttpOnly", None) ]) flow.response.cookies["UseDC"] = ("master", cookies_attribute) flow.response.cookies["UseCDNCache"] = ("false", cookies_attribute.copy()) self.cache_index += 1 # }}} Reting Flows # else: # Normal Pages {{{ # if url_key==r"/Special:RCWidget\?*": filename = "%2fSpecial:RCWidget?function=WH.RCWidget.rcwOnLoadData&GuVHo&nabrequest=0&anonview=1" else: filename = url_path.replace("/", "%2f") if len(filename)>100: filename = filename[:100] filename = os.path.join(self.replay_path, filename) #ctx.log.info("Requesting {:}".format(filename)) if not os.path.exists(filename): ctx.log.info("404: {:}".format(filename)) flow.response = http.Response.make(404) else: status_code, header, content = load_response.load_response(filename) if "content-encoding" in header\ and header["content-encoding"]==b"gzip"\ and len(content)>0: try: content = gzip.decompress(content) except Exception as e: ctx.log.info(str(e)) response_time = datetime.datetime.utcnow() #if "date" in header: #header["date"] = response_time.strftime(self.dateformat) #if "last-modified" in header: #header["last-modified"] = (response_time - datetime.timedelta(days=1))\ #.strftime(self.dateformat) #if "expires" in header: #header["expires"] = (response_time + datetime.timedelta(days=3650))\ #.strftime(self.dateformat) if "x-timer" in header: header["x-timer"] = "S{:.6f},VS0,VE0".format(response_time.timestamp()) if "x-c" in header: header["x-c"] = "cache-tyo{:d}-TYO,M".format(self.cache_index) flow.response = http.Response.make(status_code, content=content, headers=header) flow.response.refresh() #ctx.log.info("WARN: {:}".format(content==flow.response.content)) self.cache_index += 1 # }}} Normal Pages # #else: #flow.response = http.Response.make(204) #pass # }}} class `Replayer` # addons = [ Replayer( 15090 # just a positive integer is ok , "dumps" # crawled data from WikiHow website , "templates" # templates of the search result page , "indices-t/indices" # indices for the search engine based on pyserini , "indices-t/docs/doc_meta.csv" # doc meta ) ]