section stringlengths 2 30 | filename stringlengths 1 82 | text stringlengths 783 28M |
|---|---|---|
searx | webapp | #!/usr/bin/env python
"""
searx is free software: you can redistribute it and/or modify
it under the terms of the GNU Affero General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
searx is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU Affero General Public License for more details.
You should have received a copy of the GNU Affero General Public License
along with searx. If not, see < http://www.gnu.org/licenses/ >.
(C) 2013- by Adam Tauber, <asciimoo@gmail.com>
"""
import sys
if sys.version_info[0] < 3:
print("\033[1;31m Python2 is no longer supported\033[0m")
exit(1)
if __name__ == "__main__":
from os.path import dirname, realpath
sys.path.append(realpath(dirname(realpath(__file__)) + "/../"))
import hashlib
import hmac
import json
import os
import requests
from searx import logger
logger = logger.getChild("webapp")
import urllib
from datetime import datetime, timedelta
from html import escape
from io import StringIO
from time import time
from urllib.parse import urlencode, urlparse
import flask_babel
from babel.support import Translations
from flask import (
Flask,
Response,
make_response,
redirect,
render_template,
request,
send_from_directory,
url_for,
)
from flask.ctx import has_request_context
from flask.json import jsonify
from flask_babel import Babel, format_date, format_decimal, gettext
from pygments import highlight
from pygments.formatters import HtmlFormatter # pylint: disable=no-name-in-module
from pygments.lexers import get_lexer_by_name
from searx import brand, searx_debug, searx_dir, settings, static_path
from searx.answerers import answerers, ask
from searx.autocomplete import backends as autocomplete_backends
from searx.autocomplete import search_autocomplete
from searx.engines import categories, engine_shortcuts, engines, get_engines_stats
from searx.exceptions import SearxParameterException
from searx.languages import language_codes as languages
from searx.metrology.error_recorder import errors_per_engines
from searx.plugins import plugins
from searx.plugins.oa_doi_rewrite import get_doi_resolver
from searx.poolrequests import get_global_proxies
from searx.preferences import LANGUAGE_CODES, Preferences, ValidationException
from searx.query import RawTextQuery
from searx.search import SearchWithPlugins
from searx.search import initialize as search_initialize
from searx.search.checker import get_result as checker_get_result
from searx.settings_loader import get_default_settings_path
from searx.utils import dict_subset, gen_useragent, html_to_text, match_language
from searx.version import VERSION_STRING
from searx.webadapter import get_search_query_from_webapp, get_selected_categories
from searx.webutils import (
UnicodeWriter,
get_resources_directory,
get_result_templates,
get_static_files,
get_themes,
highlight_content,
is_flask_run_cmdline,
new_hmac,
prettify_url,
)
from werkzeug.middleware.proxy_fix import ProxyFix
# serve pages with HTTP/1.1
from werkzeug.serving import WSGIRequestHandler
WSGIRequestHandler.protocol_version = "HTTP/{}".format(
settings["server"].get("http_protocol_version", "1.0")
)
# check secret_key
if not searx_debug and settings["server"]["secret_key"] == "ultrasecretkey":
logger.error(
"server.secret_key is not changed. Please use something else instead of ultrasecretkey."
)
exit(1)
# about static
static_path = get_resources_directory(
searx_dir, "static", settings["ui"]["static_path"]
)
logger.debug("static directory is %s", static_path)
static_files = get_static_files(static_path)
# about templates
default_theme = settings["ui"]["default_theme"]
templates_path = get_resources_directory(
searx_dir, "templates", settings["ui"]["templates_path"]
)
logger.debug("templates directory is %s", templates_path)
themes = get_themes(templates_path)
result_templates = get_result_templates(templates_path)
global_favicons = []
for indice, theme in enumerate(themes):
global_favicons.append([])
theme_img_path = os.path.join(static_path, "themes", theme, "img", "icons")
for dirpath, dirnames, filenames in os.walk(theme_img_path):
global_favicons[indice].extend(filenames)
# Flask app
app = Flask(__name__, static_folder=static_path, template_folder=templates_path)
app.jinja_env.trim_blocks = True
app.jinja_env.lstrip_blocks = True
app.jinja_env.add_extension("jinja2.ext.loopcontrols") # pylint: disable=no-member
app.secret_key = settings["server"]["secret_key"]
# see https://flask.palletsprojects.com/en/1.1.x/cli/
# True if "FLASK_APP=searx/webapp.py FLASK_ENV=development flask run"
flask_run_development = (
os.environ.get("FLASK_APP") is not None
and os.environ.get("FLASK_ENV") == "development"
and is_flask_run_cmdline()
)
# True if reload feature is activated of werkzeug, False otherwise (including uwsgi, etc..)
# __name__ != "__main__" if searx.webapp is imported (make test, make docs, uwsgi...)
# see run() at the end of this file : searx_debug activates the reload feature.
werkzeug_reloader = flask_run_development or (searx_debug and __name__ == "__main__")
# initialize the engines except on the first run of the werkzeug server.
if not werkzeug_reloader or (
werkzeug_reloader and os.environ.get("WERKZEUG_RUN_MAIN") == "true"
):
search_initialize(enable_checker=True)
babel = Babel(app)
rtl_locales = [
"ar",
"arc",
"bcc",
"bqi",
"ckb",
"dv",
"fa",
"fa_IR",
"glk",
"he",
"ku",
"mzn",
"pnb",
"ps",
"sd",
"ug",
"ur",
"yi",
]
ui_locale_codes = [l.replace("_", "-") for l in settings["locales"].keys()]
# used when translating category names
_category_names = (
gettext("files"),
gettext("general"),
gettext("music"),
gettext("social media"),
gettext("images"),
gettext("videos"),
gettext("it"),
gettext("news"),
gettext("map"),
gettext("onions"),
gettext("science"),
)
_flask_babel_get_translations = flask_babel.get_translations
# monkey patch for flask_babel.get_translations
def _get_translations():
if has_request_context() and request.form.get("use-translation") == "oc":
babel_ext = flask_babel.current_app.extensions["babel"]
return Translations.load(next(babel_ext.translation_directories), "oc")
return _flask_babel_get_translations()
flask_babel.get_translations = _get_translations
def _get_browser_or_settings_language(request, lang_list):
for lang in request.headers.get("Accept-Language", "en").split(","):
if ";" in lang:
lang = lang.split(";")[0]
if "-" in lang:
lang_parts = lang.split("-")
lang = "{}-{}".format(lang_parts[0], lang_parts[-1].upper())
locale = match_language(lang, lang_list, fallback=None)
if locale is not None:
return locale
return settings["search"]["default_lang"] or "en"
@babel.localeselector
def get_locale():
if "locale" in request.form and request.form["locale"] in settings["locales"]:
# use locale from the form
locale = request.form["locale"]
locale_source = "form"
elif request.preferences.get_value("locale") != "":
# use locale from the preferences
locale = request.preferences.get_value("locale")
locale_source = "preferences"
else:
# use local from the browser
locale = _get_browser_or_settings_language(request, ui_locale_codes)
locale = locale.replace("-", "_")
locale_source = "browser"
# see _get_translations function
# and https://github.com/searx/searx/pull/1863
if locale == "oc":
request.form["use-translation"] = "oc"
locale = "fr_FR"
logger.debug(
"%s uses locale `%s` from %s",
urllib.parse.quote(request.url),
locale,
locale_source,
)
return locale
# code-highlighter
@app.template_filter("code_highlighter")
def code_highlighter(codelines, language=None):
if not language:
language = "text"
try:
# find lexer by programming language
lexer = get_lexer_by_name(language, stripall=True)
except:
# if lexer is not found, using default one
logger.debug("highlighter cannot find lexer for {0}".format(language))
lexer = get_lexer_by_name("text", stripall=True)
html_code = ""
tmp_code = ""
last_line = None
# parse lines
for line, code in codelines:
if not last_line:
line_code_start = line
# new codeblock is detected
if last_line is not None and last_line + 1 != line:
# highlight last codepart
formatter = HtmlFormatter(
linenos="inline", linenostart=line_code_start, cssclass="code-highlight"
)
html_code = html_code + highlight(tmp_code, lexer, formatter)
# reset conditions for next codepart
tmp_code = ""
line_code_start = line
# add codepart
tmp_code += code + "\n"
# update line
last_line = line
# highlight last codepart
formatter = HtmlFormatter(
linenos="inline", linenostart=line_code_start, cssclass="code-highlight"
)
html_code = html_code + highlight(tmp_code, lexer, formatter)
return html_code
# Extract domain from url
@app.template_filter("extract_domain")
def extract_domain(url):
return urlparse(url)[1]
def get_base_url():
return url_for("index", _external=True)
def get_current_theme_name(override=None):
"""Returns theme name.
Checks in this order:
1. override
2. cookies
3. settings"""
if override and (override in themes or override == "__common__"):
return override
theme_name = request.args.get("theme", request.preferences.get_value("theme"))
if theme_name not in themes:
theme_name = default_theme
return theme_name
def get_result_template(theme, template_name):
themed_path = theme + "/result_templates/" + template_name
if themed_path in result_templates:
return themed_path
return "result_templates/" + template_name
def url_for_theme(endpoint, override_theme=None, **values):
if endpoint == "static" and values.get("filename"):
theme_name = get_current_theme_name(override=override_theme)
filename_with_theme = "themes/{}/{}".format(theme_name, values["filename"])
if filename_with_theme in static_files:
values["filename"] = filename_with_theme
url = url_for(endpoint, **values)
return url
def proxify(url):
if url.startswith("//"):
url = "https:" + url
if not settings.get("result_proxy"):
return url
url_params = dict(mortyurl=url.encode())
if settings["result_proxy"].get("key"):
url_params["mortyhash"] = hmac.new(
settings["result_proxy"]["key"], url.encode(), hashlib.sha256
).hexdigest()
return "{0}?{1}".format(settings["result_proxy"]["url"], urlencode(url_params))
def image_proxify(url):
if url.startswith("//"):
url = "https:" + url
if not request.preferences.get_value("image_proxy"):
return url
if url.startswith("data:image/"):
# 50 is an arbitrary number to get only the beginning of the image.
partial_base64 = url[len("data:image/") : 50].split(";")
if (
len(partial_base64) == 2
and partial_base64[0]
in ["gif", "png", "jpeg", "pjpeg", "webp", "tiff", "bmp"]
and partial_base64[1].startswith("base64,")
):
return url
else:
return None
if settings.get("result_proxy"):
return proxify(url)
h = new_hmac(settings["server"]["secret_key"], url.encode())
return "{0}?{1}".format(
url_for("image_proxy"), urlencode(dict(url=url.encode(), h=h))
)
def get_translations():
return {
# when there is autocompletion
"no_item_found": gettext("No item found")
}
def render(template_name, override_theme=None, **kwargs):
disabled_engines = request.preferences.engines.get_disabled()
enabled_categories = set(
category
for engine_name in engines
for category in engines[engine_name].categories
if (engine_name, category) not in disabled_engines
)
if "categories" not in kwargs:
kwargs["categories"] = [
x for x in _get_ordered_categories() if x in enabled_categories
]
if "autocomplete" not in kwargs:
kwargs["autocomplete"] = request.preferences.get_value("autocomplete")
locale = request.preferences.get_value("locale")
if locale in rtl_locales and "rtl" not in kwargs:
kwargs["rtl"] = True
kwargs["searx_version"] = VERSION_STRING
kwargs["method"] = request.preferences.get_value("method")
kwargs["autofocus"] = request.preferences.get_value("autofocus")
kwargs["archive_today"] = request.preferences.get_value("archive_today")
kwargs["safesearch"] = str(request.preferences.get_value("safesearch"))
kwargs["language_codes"] = languages
if "current_language" not in kwargs:
kwargs["current_language"] = match_language(
request.preferences.get_value("language"), LANGUAGE_CODES
)
# override url_for function in templates
kwargs["url_for"] = url_for_theme
kwargs["image_proxify"] = image_proxify
kwargs["proxify"] = proxify if settings.get("result_proxy", {}).get("url") else None
kwargs["opensearch_url"] = (
url_for("opensearch")
+ "?"
+ urlencode(
{"method": kwargs["method"], "autocomplete": kwargs["autocomplete"]}
)
)
kwargs["get_result_template"] = get_result_template
kwargs["theme"] = get_current_theme_name(override=override_theme)
kwargs["template_name"] = template_name
kwargs["cookies"] = request.cookies
kwargs["errors"] = request.errors
kwargs["instance_name"] = settings["general"]["instance_name"]
kwargs["results_on_new_tab"] = request.preferences.get_value("results_on_new_tab")
kwargs["preferences"] = request.preferences
kwargs["brand"] = brand
kwargs["translations"] = json.dumps(get_translations(), separators=(",", ":"))
kwargs["scripts"] = set()
kwargs["endpoint"] = "results" if "q" in kwargs else request.endpoint
for plugin in request.user_plugins:
for script in plugin.js_dependencies:
kwargs["scripts"].add(script)
kwargs["styles"] = set()
for plugin in request.user_plugins:
for css in plugin.css_dependencies:
kwargs["styles"].add(css)
return render_template("{}/{}".format(kwargs["theme"], template_name), **kwargs)
def _get_ordered_categories():
ordered_categories = []
if "categories_order" not in settings["ui"]:
ordered_categories = ["general"]
ordered_categories.extend(
x for x in sorted(categories.keys()) if x != "general"
)
return ordered_categories
ordered_categories = settings["ui"]["categories_order"]
ordered_categories.extend(
x for x in sorted(categories.keys()) if x not in ordered_categories
)
return ordered_categories
@app.before_request
def pre_request():
request.start_time = time()
request.timings = []
request.errors = []
preferences = Preferences(themes, list(categories.keys()), engines, plugins)
user_agent = request.headers.get("User-Agent", "").lower()
if "webkit" in user_agent and "android" in user_agent:
preferences.key_value_settings["method"].value = "GET"
request.preferences = preferences
try:
preferences.parse_dict(request.cookies)
except:
request.errors.append(gettext("Invalid settings, please edit your preferences"))
# merge GET, POST vars
# request.form
request.form = dict(request.form.items())
for k, v in request.args.items():
if k not in request.form:
request.form[k] = v
if request.form.get("preferences"):
preferences.parse_encoded_data(request.form["preferences"])
else:
try:
preferences.parse_dict(request.form)
except Exception:
logger.exception("invalid settings")
request.errors.append(gettext("Invalid settings"))
# init search language and locale
if not preferences.get_value("language"):
preferences.parse_dict(
{"language": _get_browser_or_settings_language(request, LANGUAGE_CODES)}
)
if not preferences.get_value("locale"):
preferences.parse_dict({"locale": get_locale()})
# request.user_plugins
request.user_plugins = []
allowed_plugins = preferences.plugins.get_enabled()
disabled_plugins = preferences.plugins.get_disabled()
for plugin in plugins:
if (
plugin.default_on and plugin.id not in disabled_plugins
) or plugin.id in allowed_plugins:
request.user_plugins.append(plugin)
@app.after_request
def add_default_headers(response):
# set default http headers
for header, value in settings["server"].get("default_http_headers", {}).items():
if header in response.headers:
continue
response.headers[header] = value
return response
@app.after_request
def post_request(response):
total_time = time() - request.start_time
timings_all = ["total;dur=" + str(round(total_time * 1000, 3))]
if len(request.timings) > 0:
timings = sorted(request.timings, key=lambda v: v["total"])
timings_total = [
"total_"
+ str(i)
+ "_"
+ v["engine"]
+ ";dur="
+ str(round(v["total"] * 1000, 3))
for i, v in enumerate(timings)
]
timings_load = [
"load_"
+ str(i)
+ "_"
+ v["engine"]
+ ";dur="
+ str(round(v["load"] * 1000, 3))
for i, v in enumerate(timings)
]
timings_all = timings_all + timings_total + timings_load
response.headers.add("Server-Timing", ", ".join(timings_all))
return response
def index_error(output_format, error_message):
if output_format == "json":
return Response(
json.dumps({"error": error_message}), mimetype="application/json"
)
elif output_format == "csv":
response = Response("", mimetype="application/csv")
cont_disp = "attachment;Filename=searx.csv"
response.headers.add("Content-Disposition", cont_disp)
return response
elif output_format == "rss":
response_rss = render(
"opensearch_response_rss.xml",
results=[],
q=request.form["q"] if "q" in request.form else "",
number_of_results=0,
base_url=get_base_url(),
error_message=error_message,
override_theme="__common__",
)
return Response(response_rss, mimetype="text/xml")
else:
# html
request.errors.append(gettext("search error"))
return render(
"index.html",
selected_categories=get_selected_categories(
request.preferences, request.form
),
)
@app.route("/", methods=["GET", "POST"])
def index():
"""Render index page."""
# UI
advanced_search = request.preferences.get_value("advanced_search")
# redirect to search if there's a query in the request
if request.form.get("q"):
query = ("?" + request.query_string.decode()) if request.query_string else ""
return redirect(url_for("search") + query, 308)
return render(
"index.html",
selected_categories=get_selected_categories(request.preferences, request.form),
advanced_search=advanced_search,
)
@app.route("/healthz", methods=["GET"])
def health():
return Response("OK", mimetype="text/plain")
@app.route("/search", methods=["GET", "POST"])
def search():
"""Search query in q and return results.
Supported outputs: html, json, csv, rss.
"""
# output_format
output_format = request.form.get("format", "html")
if output_format not in ["html", "csv", "json", "rss"]:
output_format = "html"
# check if there is query (not None and not an empty string)
if not request.form.get("q"):
if output_format == "html":
return render(
"index.html",
advanced_search=request.preferences.get_value("advanced_search"),
selected_categories=get_selected_categories(
request.preferences, request.form
),
)
else:
return index_error(output_format, "No query"), 400
# search
search_query = None
raw_text_query = None
result_container = None
try:
search_query, raw_text_query, _, _ = get_search_query_from_webapp(
request.preferences, request.form
)
# search = Search(search_query) # without plugins
search = SearchWithPlugins(search_query, request.user_plugins, request)
result_container = search.search()
except SearxParameterException as e:
logger.exception("search error: SearxParameterException")
return index_error(output_format, e.message), 400
except Exception as e:
logger.exception("search error")
return index_error(output_format, gettext("search error")), 500
# results
results = result_container.get_ordered_results()
number_of_results = result_container.results_number()
if number_of_results < result_container.results_length():
number_of_results = 0
# checkin for a external bang
if result_container.redirect_url:
return redirect(result_container.redirect_url)
# Server-Timing header
request.timings = result_container.get_timings()
# output
for result in results:
if output_format == "html":
if "content" in result and result["content"]:
result["content"] = highlight_content(
escape(result["content"][:1024]), search_query.query
)
if "title" in result and result["title"]:
result["title"] = highlight_content(
escape(result["title"] or ""), search_query.query
)
else:
if result.get("content"):
result["content"] = html_to_text(result["content"]).strip()
# removing html content and whitespace duplications
result["title"] = " ".join(html_to_text(result["title"]).strip().split())
if "url" in result and "pretty_url" not in result:
result["pretty_url"] = prettify_url(result["url"])
# TODO, check if timezone is calculated right
if result.get(
"publishedDate"
): # do not try to get a date from an empty string or a None type
try: # test if publishedDate >= 1900 (datetime module bug)
result["pubdate"] = result["publishedDate"].strftime(
"%Y-%m-%d %H:%M:%S%z"
)
except ValueError:
result["publishedDate"] = None
else:
if result["publishedDate"].replace(
tzinfo=None
) >= datetime.now() - timedelta(days=1):
timedifference = datetime.now() - result["publishedDate"].replace(
tzinfo=None
)
minutes = int((timedifference.seconds / 60) % 60)
hours = int(timedifference.seconds / 60 / 60)
if hours == 0:
result["publishedDate"] = gettext(
"{minutes} minute(s) ago"
).format(minutes=minutes)
else:
result["publishedDate"] = gettext(
"{hours} hour(s), {minutes} minute(s) ago"
).format(hours=hours, minutes=minutes) # noqa
else:
result["publishedDate"] = format_date(result["publishedDate"])
if output_format == "json":
return Response(
json.dumps(
{
"query": search_query.query,
"number_of_results": number_of_results,
"results": results,
"answers": list(result_container.answers),
"corrections": list(result_container.corrections),
"infoboxes": result_container.infoboxes,
"suggestions": list(result_container.suggestions),
"unresponsive_engines": __get_translated_errors(
result_container.unresponsive_engines
),
}, # noqa
default=lambda item: list(item) if isinstance(item, set) else item,
),
mimetype="application/json",
)
elif output_format == "csv":
csv = UnicodeWriter(StringIO())
keys = ("title", "url", "content", "host", "engine", "score", "type")
csv.writerow(keys)
for row in results:
row["host"] = row["parsed_url"].netloc
row["type"] = "result"
csv.writerow([row.get(key, "") for key in keys])
for a in result_container.answers:
row = {"title": a, "type": "answer"}
csv.writerow([row.get(key, "") for key in keys])
for a in result_container.suggestions:
row = {"title": a, "type": "suggestion"}
csv.writerow([row.get(key, "") for key in keys])
for a in result_container.corrections:
row = {"title": a, "type": "correction"}
csv.writerow([row.get(key, "") for key in keys])
csv.stream.seek(0)
response = Response(csv.stream.read(), mimetype="application/csv")
cont_disp = "attachment;Filename=searx_-_{0}.csv".format(search_query.query)
response.headers.add("Content-Disposition", cont_disp)
return response
elif output_format == "rss":
response_rss = render(
"opensearch_response_rss.xml",
results=results,
answers=result_container.answers,
corrections=result_container.corrections,
suggestions=result_container.suggestions,
q=request.form["q"],
number_of_results=number_of_results,
base_url=get_base_url(),
override_theme="__common__",
)
return Response(response_rss, mimetype="text/xml")
# HTML output format
# suggestions: use RawTextQuery to get the suggestion URLs with the same bang
suggestion_urls = list(
map(
lambda suggestion: {
"url": raw_text_query.changeQuery(suggestion).getFullQuery(),
"title": suggestion,
},
result_container.suggestions,
)
)
correction_urls = list(
map(
lambda correction: {
"url": raw_text_query.changeQuery(correction).getFullQuery(),
"title": correction,
},
result_container.corrections,
)
)
#
return render(
"results.html",
results=results,
q=request.form["q"],
selected_categories=search_query.categories,
pageno=search_query.pageno,
time_range=search_query.time_range,
number_of_results=format_decimal(number_of_results),
suggestions=suggestion_urls,
answers=result_container.answers,
corrections=correction_urls,
infoboxes=result_container.infoboxes,
engine_data=result_container.engine_data,
paging=result_container.paging,
unresponsive_engines=__get_translated_errors(
result_container.unresponsive_engines
),
current_language=match_language(
search_query.lang,
LANGUAGE_CODES,
fallback=request.preferences.get_value("language"),
),
base_url=get_base_url(),
theme=get_current_theme_name(),
favicons=global_favicons[themes.index(get_current_theme_name())],
timeout_limit=request.form.get("timeout_limit", None),
)
def __get_translated_errors(unresponsive_engines):
translated_errors = set()
for unresponsive_engine in unresponsive_engines:
error_msg = gettext(unresponsive_engine[1])
if unresponsive_engine[2]:
error_msg = "{} {}".format(error_msg, unresponsive_engine[2])
translated_errors.add((unresponsive_engine[0], error_msg))
return translated_errors
@app.route("/about", methods=["GET"])
def about():
"""Render about page"""
return render(
"about.html",
)
@app.route("/autocompleter", methods=["GET", "POST"])
def autocompleter():
"""Return autocompleter results"""
# run autocompleter
results = []
# set blocked engines
disabled_engines = request.preferences.engines.get_disabled()
# parse query
raw_text_query = RawTextQuery(request.form.get("q", ""), disabled_engines)
sug_prefix = raw_text_query.getQuery()
# normal autocompletion results only appear if no inner results returned
# and there is a query part
if len(raw_text_query.autocomplete_list) == 0 and len(sug_prefix) > 0:
# get language from cookie
language = request.preferences.get_value("language")
if not language or language == "all":
language = "en"
else:
language = language.split("-")[0]
# run autocompletion
raw_results = search_autocomplete(
request.preferences.get_value("autocomplete"), sug_prefix, language
)
for result in raw_results:
# attention: this loop will change raw_text_query object and this is
# the reason why the sug_prefix was stored before (see above)
results.append(raw_text_query.changeQuery(result).getFullQuery())
if len(raw_text_query.autocomplete_list) > 0:
for autocomplete_text in raw_text_query.autocomplete_list:
results.append(
raw_text_query.get_autocomplete_full_query(autocomplete_text)
)
for answers in ask(raw_text_query):
for answer in answers:
results.append(str(answer["answer"]))
if request.headers.get("X-Requested-With") == "XMLHttpRequest":
# the suggestion request comes from the searx search form
suggestions = json.dumps(results)
mimetype = "application/json"
else:
# the suggestion request comes from browser's URL bar
suggestions = json.dumps([sug_prefix, results])
mimetype = "application/x-suggestions+json"
return Response(suggestions, mimetype=mimetype)
@app.route("/preferences", methods=["GET", "POST"])
def preferences():
"""Render preferences page && save user preferences"""
# save preferences
if request.method == "POST":
resp = make_response(redirect(url_for("index", _external=True)))
try:
request.preferences.parse_form(request.form)
except ValidationException:
request.errors.append(
gettext("Invalid settings, please edit your preferences")
)
return resp
return request.preferences.save(resp)
# render preferences
image_proxy = request.preferences.get_value("image_proxy")
disabled_engines = request.preferences.engines.get_disabled()
allowed_plugins = request.preferences.plugins.get_enabled()
# stats for preferences page
stats = {}
engines_by_category = {}
for c in categories:
engines_by_category[c] = []
for e in categories[c]:
if not request.preferences.validate_token(e):
continue
stats[e.name] = {"time": None, "warn_timeout": False, "warn_time": False}
if e.timeout > settings["outgoing"]["request_timeout"]:
stats[e.name]["warn_timeout"] = True
stats[e.name][
"supports_selected_language"
] = _is_selected_language_supported(e, request.preferences)
engines_by_category[c].append(e)
# get first element [0], the engine time,
# and then the second element [1] : the time (the first one is the label)
for engine_stat in get_engines_stats(request.preferences)[0][1]:
stats[engine_stat.get("name")]["time"] = round(engine_stat.get("avg"), 3)
if engine_stat.get("avg") > settings["outgoing"]["request_timeout"]:
stats[engine_stat.get("name")]["warn_time"] = True
# end of stats
locked_preferences = list()
if "preferences" in settings and "lock" in settings["preferences"]:
locked_preferences = settings["preferences"]["lock"]
return render(
"preferences.html",
selected_categories=get_selected_categories(request.preferences, request.form),
all_categories=_get_ordered_categories(),
locales=settings["locales"],
current_locale=request.preferences.get_value("locale"),
image_proxy=image_proxy,
engines_by_category=engines_by_category,
stats=stats,
answerers=[{"info": a.self_info(), "keywords": a.keywords} for a in answerers],
disabled_engines=disabled_engines,
autocomplete_backends=autocomplete_backends,
shortcuts={y: x for x, y in engine_shortcuts.items()},
themes=themes,
plugins=plugins,
doi_resolvers=settings["doi_resolvers"],
current_doi_resolver=get_doi_resolver(
request.args, request.preferences.get_value("doi_resolver")
),
allowed_plugins=allowed_plugins,
theme=get_current_theme_name(),
preferences_url_params=request.preferences.get_as_url_params(),
base_url=get_base_url(),
locked_preferences=locked_preferences,
preferences=True,
)
def _is_selected_language_supported(engine, preferences):
language = preferences.get_value("language")
return language == "all" or match_language(
language,
getattr(engine, "supported_languages", []),
getattr(engine, "language_aliases", {}),
None,
)
@app.route("/image_proxy", methods=["GET"])
def image_proxy():
url = request.args.get("url").encode()
if not url:
return "", 400
h = new_hmac(settings["server"]["secret_key"], url)
if h != request.args.get("h"):
return "", 400
headers = {
"User-Agent": gen_useragent(),
"Accept": "image/webp,*/*",
"Accept-Encoding": "gzip, deflate",
"Sec-GPC": "1",
"DNT": "1",
}
headers = dict_subset(request.headers, {"If-Modified-Since", "If-None-Match"})
resp = requests.get(
url,
stream=True,
timeout=settings["outgoing"]["request_timeout"],
headers=headers,
proxies=get_global_proxies(),
)
if resp.status_code == 304:
return "", resp.status_code
if resp.status_code != 200:
logger.debug("image-proxy: wrong response code: {0}".format(resp.status_code))
if resp.status_code >= 400:
return "", resp.status_code
return "", 400
if not resp.headers.get("content-type", "").startswith("image/"):
logger.debug(
"image-proxy: wrong content-type: {0}".format(
resp.headers.get("content-type")
)
)
return "", 400
img = b""
chunk_counter = 0
for chunk in resp.iter_content(1024 * 1024):
chunk_counter += 1
if chunk_counter > 5:
return "", 502 # Bad gateway - file is too big (>5M)
img += chunk
headers = dict_subset(
resp.headers,
{"Content-Length", "Length", "Date", "Last-Modified", "Expires", "Etag"},
)
return Response(img, mimetype=resp.headers["content-type"], headers=headers)
@app.route("/stats", methods=["GET"])
def stats():
"""Render engine statistics page."""
if not settings["general"].get("enable_stats"):
return page_not_found(None)
stats = get_engines_stats(request.preferences)
return render(
"stats.html",
stats=stats,
)
@app.route("/stats/errors", methods=["GET"])
def stats_errors():
result = {}
engine_names = list(errors_per_engines.keys())
engine_names.sort()
for engine_name in engine_names:
error_stats = errors_per_engines[engine_name]
sent_search_count = max(engines[engine_name].stats["sent_search_count"], 1)
sorted_context_count_list = sorted(
error_stats.items(), key=lambda context_count: context_count[1]
)
r = []
percentage_sum = 0
for context, count in sorted_context_count_list:
percentage = round(20 * count / sent_search_count) * 5
percentage_sum += percentage
r.append(
{
"filename": context.filename,
"function": context.function,
"line_no": context.line_no,
"code": context.code,
"exception_classname": context.exception_classname,
"log_message": context.log_message,
"log_parameters": context.log_parameters,
"percentage": percentage,
}
)
result[engine_name] = sorted(r, reverse=True, key=lambda d: d["percentage"])
return jsonify(result)
@app.route("/stats/checker", methods=["GET"])
def stats_checker():
result = checker_get_result()
return jsonify(result)
@app.route("/robots.txt", methods=["GET"])
def robots():
return Response(
"""User-agent: *
Allow: /
Allow: /about
Disallow: /stats
Disallow: /preferences
Disallow: /*?*q=*
""",
mimetype="text/plain",
)
@app.route("/opensearch.xml", methods=["GET"])
def opensearch():
method = "post"
if request.preferences.get_value("method") == "GET":
method = "get"
# chrome/chromium only supports HTTP GET....
if request.headers.get("User-Agent", "").lower().find("webkit") >= 0:
method = "get"
ret = render(
"opensearch.xml", opensearch_method=method, override_theme="__common__"
)
resp = Response(
response=ret, status=200, mimetype="application/opensearchdescription+xml"
)
return resp
@app.route("/favicon.ico")
def favicon():
return send_from_directory(
os.path.join(
app.root_path, static_path, "themes", get_current_theme_name(), "img"
),
"favicon.png",
mimetype="image/vnd.microsoft.icon",
)
@app.route("/clear_cookies")
def clear_cookies():
resp = make_response(redirect(url_for("index", _external=True)))
for cookie_name in request.cookies:
resp.delete_cookie(cookie_name)
return resp
@app.route("/config")
def config():
"""Return configuration in JSON format."""
_engines = []
for name, engine in engines.items():
if not request.preferences.validate_token(engine):
continue
supported_languages = engine.supported_languages
if isinstance(engine.supported_languages, dict):
supported_languages = list(engine.supported_languages.keys())
_engines.append(
{
"name": name,
"categories": engine.categories,
"shortcut": engine.shortcut,
"enabled": not engine.disabled,
"paging": engine.paging,
"language_support": engine.language_support,
"supported_languages": supported_languages,
"safesearch": engine.safesearch,
"time_range_support": engine.time_range_support,
"timeout": engine.timeout,
}
)
_plugins = []
for _ in plugins:
_plugins.append({"name": _.name, "enabled": _.default_on})
return jsonify(
{
"categories": list(categories.keys()),
"engines": _engines,
"plugins": _plugins,
"instance_name": settings["general"]["instance_name"],
"locales": settings["locales"],
"default_locale": settings["ui"]["default_locale"],
"autocomplete": settings["search"]["autocomplete"],
"safe_search": settings["search"]["safe_search"],
"default_theme": settings["ui"]["default_theme"],
"version": VERSION_STRING,
"brand": {
"CONTACT_URL": brand.CONTACT_URL,
"GIT_URL": brand.GIT_URL,
"GIT_BRANCH": brand.GIT_BRANCH,
"DOCS_URL": brand.DOCS_URL,
},
"doi_resolvers": [r for r in settings["doi_resolvers"]],
"default_doi_resolver": settings["default_doi_resolver"],
}
)
@app.errorhandler(404)
def page_not_found(e):
return render("404.html"), 404
def run():
logger.debug(
"starting webserver on %s:%s",
settings["server"]["bind_address"],
settings["server"]["port"],
)
app.run(
debug=searx_debug,
use_debugger=searx_debug,
port=settings["server"]["port"],
host=settings["server"]["bind_address"],
threaded=True,
extra_files=[get_default_settings_path()],
)
def patch_application(app):
# serve pages with HTTP/1.1
WSGIRequestHandler.protocol_version = "HTTP/{}".format(
settings["server"]["http_protocol_version"]
)
# patch app to handle non root url-s behind proxy & wsgi
app.wsgi_app = ReverseProxyPathFix(ProxyFix(app.wsgi_app))
class ReverseProxyPathFix:
"""Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
http://flask.pocoo.org/snippets/35/
In nginx:
location /myprefix {
proxy_pass http://127.0.0.1:8000;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
"""
def __init__(self, app):
self.app = app
self.script_name = None
self.scheme = None
self.server = None
if settings["server"]["base_url"]:
# If base_url is specified, then these values from are given
# preference over any Flask's generics.
base_url = urlparse(settings["server"]["base_url"])
self.script_name = base_url.path
if self.script_name.endswith("/"):
# remove trailing slash to avoid infinite redirect on the index
# see https://github.com/searx/searx/issues/2729
self.script_name = self.script_name[:-1]
self.scheme = base_url.scheme
self.server = base_url.netloc
def __call__(self, environ, start_response):
script_name = self.script_name or environ.get("HTTP_X_SCRIPT_NAME", "")
if script_name:
environ["SCRIPT_NAME"] = script_name
path_info = environ["PATH_INFO"]
if path_info.startswith(script_name):
environ["PATH_INFO"] = path_info[len(script_name) :]
scheme = self.scheme or environ.get("HTTP_X_SCHEME", "")
if scheme:
environ["wsgi.url_scheme"] = scheme
server = self.server or environ.get("HTTP_X_FORWARDED_HOST", "")
if server:
environ["HTTP_HOST"] = server
return self.app(environ, start_response)
application = app
# patch app to handle non root url-s behind proxy & wsgi
app.wsgi_app = ReverseProxyPathFix(ProxyFix(application.wsgi_app))
if __name__ == "__main__":
run()
|
ui | ratingwidget | # -*- coding: utf-8 -*-
#
# Picard, the next-generation MusicBrainz tagger
#
# Copyright (C) 2008, 2018-2022 Philipp Wolfer
# Copyright (C) 2011, 2013 Michael Wiencek
# Copyright (C) 2013, 2018, 2020-2022 Laurent Monin
# Copyright (C) 2016-2017 Sambhav Kothari
# Copyright (C) 2018 Vishal Choudhary
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
from picard import log
from picard.config import get_config
from PyQt5 import QtCore, QtGui, QtWidgets
class RatingWidget(QtWidgets.QWidget):
def __init__(self, parent, track):
super().__init__(parent)
self._track = track
config = get_config()
self._maximum = config.setting["rating_steps"] - 1
try:
self._rating = int(track.metadata["~rating"] or 0)
except ValueError:
self._rating = 0
self._highlight = 0
self._star_pixmap = QtGui.QPixmap(":/images/star.png")
self._star_gray_pixmap = QtGui.QPixmap(":/images/star-gray.png")
self._star_size = 16
self._star_spacing = 2
self._offset = 16
self._width = (
self._maximum * (self._star_size + self._star_spacing) + self._offset
)
self._height = self._star_size + 6
self.setMaximumSize(self._width, self._height)
self.setMinimumSize(self._width, self._height)
self.setSizePolicy(
QtWidgets.QSizePolicy(
QtWidgets.QSizePolicy.Policy.Fixed, QtWidgets.QSizePolicy.Policy.Fixed
)
)
self.setMouseTracking(True)
def sizeHint(self):
return QtCore.QSize(self._width, self._height)
def _setHighlight(self, highlight):
assert 0 <= highlight <= self._maximum
if highlight != self._highlight:
self._highlight = highlight
self.update()
def mousePressEvent(self, event):
if event.button() == QtCore.Qt.MouseButton.LeftButton:
x = event.x()
if x < self._offset:
return
rating = self._getRatingFromPosition(x)
if self._rating == rating:
rating = 0
self._rating = rating
self._update_track()
self.update()
event.accept()
def mouseMoveEvent(self, event):
self._setHighlight(self._getRatingFromPosition(event.x()))
event.accept()
def leaveEvent(self, event):
self._setHighlight(0)
event.accept()
def _getRatingFromPosition(self, position):
rating = (
int((position - self._offset) / (self._star_size + self._star_spacing)) + 1
)
if rating > self._maximum:
rating = self._maximum
return rating
def _submitted(self, document, http, error):
if error:
self.tagger.window.set_statusbar_message(
N_(
"Failed to submit rating for track '%(track_title)s' due to server error %(error)d"
),
{"track_title": self._track.metadata["title"], "error": error},
echo=None,
)
log.error(
"Failed to submit rating for %s (server HTTP error %d)",
self._track,
error,
)
def _update_track(self):
track = self._track
rating = str(self._rating)
track.metadata["~rating"] = rating
for file in track.files:
file.metadata["~rating"] = rating
config = get_config()
if config.setting["submit_ratings"]:
ratings = {("recording", track.id): self._rating}
try:
self.tagger.mb_api.submit_ratings(ratings, self._submitted)
except (
ValueError
): # This should never happen as self._rating is always an integer
log.error(
"Failed to submit rating for recording %s", track.id, exc_info=True
)
def paintEvent(self, event=None):
painter = QtGui.QPainter(self)
offset = self._offset
for i in range(1, self._maximum + 1):
if i <= self._rating or i <= self._highlight:
pixmap = self._star_pixmap
else:
pixmap = self._star_gray_pixmap
painter.drawPixmap(offset, 3, pixmap)
offset += self._star_size + self._star_spacing
|
notifier | telegram | from __future__ import absolute_import
import asyncio
import logging
from ..conf import settings
logger = logging.getLogger(__name__)
class TelegramBot:
def __init__(self, chat_id=None, split_on=None):
from telegram import Bot # pylint: disable=import-outside-toplevel
telegram_creds = settings().creds["telegram"]
token = telegram_creds["token"]
if chat_id is not None:
self._chat_id = chat_id
else:
self._chat_id = telegram_creds.get("chat")
self.split_on = split_on
self.bot = Bot(token=token)
@property
def chat_id(self):
if self._chat_id is None:
chat = self.bot.getUpdates(limit=1)[0].message.chat
logger.debug("Imprinted chat id %d of type %s", chat.id, chat.type)
self._chat_id = chat.id
return self._chat_id
def post(self, report, **kwargs):
del kwargs
if self.split_on:
report = report.split(self.split_on)
else:
report = [report]
for r in report:
# Telegram max message length is 4096 chars
messages = [r[i : i + 4096] for i in range(0, len(r), 4096)]
for m in messages:
self.send_message(m)
def send_message(self, message):
message = self.bot.send_message(
self.chat_id,
message,
parse_mode="Markdown",
)
asyncio.run(message)
return message
__call__ = post
def notify_factory(conf, value):
del conf
try:
chat_id = value["chat"]
except (TypeError, KeyError):
chat_id = value
try:
split_on = value["split-on"]
except (TypeError, KeyError):
split_on = None
return TelegramBot(chat_id=chat_id, split_on=split_on).post
def get_chat_id():
bot = TelegramBot()
print(bot.chat_id)
|
wsgiserver | wsgiserver2 | """A high-speed, production ready, thread pooled, generic HTTP server.
Simplest example on how to use this module directly
(without using CherryPy's application machinery)::
from cherrypy import wsgiserver
def my_crazy_app(environ, start_response):
status = '200 OK'
response_headers = [('Content-type','text/plain')]
start_response(status, response_headers)
return ['Hello world!']
server = wsgiserver.CherryPyWSGIServer(
('0.0.0.0', 8070), my_crazy_app,
server_name='www.cherrypy.example')
server.start()
The CherryPy WSGI server can serve as many WSGI applications
as you want in one instance by using a WSGIPathInfoDispatcher::
d = WSGIPathInfoDispatcher({'/': my_crazy_app, '/blog': my_blog_app})
server = wsgiserver.CherryPyWSGIServer(('0.0.0.0', 80), d)
Want SSL support? Just set server.ssl_adapter to an SSLAdapter instance.
This won't call the CherryPy engine (application side) at all, only the
HTTP server, which is independent from the rest of CherryPy. Don't
let the name "CherryPyWSGIServer" throw you; the name merely reflects
its origin, not its coupling.
For those of you wanting to understand internals of this module, here's the
basic call flow. The server's listening thread runs a very tight loop,
sticking incoming connections onto a Queue::
server = CherryPyWSGIServer(...)
server.start()
while True:
tick()
# This blocks until a request comes in:
child = socket.accept()
conn = HTTPConnection(child, ...)
server.requests.put(conn)
Worker threads are kept in a pool and poll the Queue, popping off and then
handling each connection in turn. Each connection can consist of an arbitrary
number of requests and their responses, so we run a nested loop::
while True:
conn = server.requests.get()
conn.communicate()
-> while True:
req = HTTPRequest(...)
req.parse_request()
-> # Read the Request-Line, e.g. "GET /page HTTP/1.1"
req.rfile.readline()
read_headers(req.rfile, req.inheaders)
req.respond()
-> response = app(...)
try:
for chunk in response:
if chunk:
req.write(chunk)
finally:
if hasattr(response, "close"):
response.close()
if req.close_connection:
return
"""
__all__ = ['HTTPRequest', 'HTTPConnection', 'HTTPServer',
'SizeCheckWrapper', 'KnownLengthRFile', 'ChunkedRFile',
'CP_fileobject',
'MaxSizeExceeded', 'NoSSLError', 'FatalSSLAlert',
'WorkerThread', 'ThreadPool', 'SSLAdapter',
'CherryPyWSGIServer',
'Gateway', 'WSGIGateway', 'WSGIGateway_10', 'WSGIGateway_u0',
'WSGIPathInfoDispatcher', 'get_ssl_adapter_class']
import os
try:
import queue
except:
import Queue as queue
import re
import socket
import sys
import rfc822
if 'win' in sys.platform and hasattr(socket, "AF_INET6"):
if not hasattr(socket, 'IPPROTO_IPV6'):
socket.IPPROTO_IPV6 = 41
if not hasattr(socket, 'IPV6_V6ONLY'):
socket.IPV6_V6ONLY = 27
try:
import cStringIO as StringIO
except ImportError:
import StringIO
DEFAULT_BUFFER_SIZE = -1
class FauxSocket(object):
"""Faux socket with the minimal interface required by pypy"""
def _reuse(self):
pass
_fileobject_uses_str_type = isinstance(
socket._fileobject(FauxSocket())._rbuf, basestring)
del FauxSocket # this class is not longer required for anything.
import threading
import time
import traceback
def format_exc(limit=None):
"""Like print_exc() but return a string. Backport for Python 2.3."""
try:
etype, value, tb = sys.exc_info()
return ''.join(traceback.format_exception(etype, value, tb, limit))
finally:
etype = value = tb = None
import operator
import warnings
from urllib import unquote
if sys.version_info >= (3, 0):
bytestr = bytes
unicodestr = str
basestring = (bytes, str)
def ntob(n, encoding='ISO-8859-1'):
"""Return the given native string as a byte string in the given
encoding.
"""
# In Python 3, the native string type is unicode
return n.encode(encoding)
else:
bytestr = str
unicodestr = unicode
basestring = basestring
def ntob(n, encoding='ISO-8859-1'):
"""Return the given native string as a byte string in the given
encoding.
"""
# In Python 2, the native string type is bytes. Assume it's already
# in the given encoding, which for ISO-8859-1 is almost always what
# was intended.
return n
LF = ntob('\n')
CRLF = ntob('\r\n')
TAB = ntob('\t')
SPACE = ntob(' ')
COLON = ntob(':')
SEMICOLON = ntob(';')
EMPTY = ntob('')
NUMBER_SIGN = ntob('#')
QUESTION_MARK = ntob('?')
ASTERISK = ntob('*')
FORWARD_SLASH = ntob('/')
quoted_slash = re.compile(ntob("(?i)%2F"))
import errno
def plat_specific_errors(*errnames):
"""Return error numbers for all errors in errnames on this platform.
The 'errno' module contains different global constants depending on
the specific platform (OS). This function will return the list of
numeric values for a given list of potential names.
"""
errno_names = dir(errno)
nums = [getattr(errno, k) for k in errnames if k in errno_names]
# de-dupe the list
return list(dict.fromkeys(nums).keys())
socket_error_eintr = plat_specific_errors("EINTR", "WSAEINTR")
socket_errors_to_ignore = plat_specific_errors(
"EPIPE",
"EBADF", "WSAEBADF",
"ENOTSOCK", "WSAENOTSOCK",
"ETIMEDOUT", "WSAETIMEDOUT",
"ECONNREFUSED", "WSAECONNREFUSED",
"ECONNRESET", "WSAECONNRESET",
"ECONNABORTED", "WSAECONNABORTED",
"ENETRESET", "WSAENETRESET",
"EHOSTDOWN", "EHOSTUNREACH",
)
socket_errors_to_ignore.append("timed out")
socket_errors_to_ignore.append("The read operation timed out")
socket_errors_nonblocking = plat_specific_errors(
'EAGAIN', 'EWOULDBLOCK', 'WSAEWOULDBLOCK')
comma_separated_headers = [
ntob(h) for h in
['Accept', 'Accept-Charset', 'Accept-Encoding',
'Accept-Language', 'Accept-Ranges', 'Allow', 'Cache-Control',
'Connection', 'Content-Encoding', 'Content-Language', 'Expect',
'If-Match', 'If-None-Match', 'Pragma', 'Proxy-Authenticate', 'TE',
'Trailer', 'Transfer-Encoding', 'Upgrade', 'Vary', 'Via', 'Warning',
'WWW-Authenticate']
]
import logging
if not hasattr(logging, 'statistics'):
logging.statistics = {}
def read_headers(rfile, hdict=None):
"""Read headers from the given stream into the given header dict.
If hdict is None, a new header dict is created. Returns the populated
header dict.
Headers which are repeated are folded together using a comma if their
specification so dictates.
This function raises ValueError when the read bytes violate the HTTP spec.
You should probably return "400 Bad Request" if this happens.
"""
if hdict is None:
hdict = {}
while True:
line = rfile.readline()
if not line:
# No more data--illegal end of headers
raise ValueError("Illegal end of headers.")
if line == CRLF:
# Normal end of headers
break
if not line.endswith(CRLF):
raise ValueError("HTTP requires CRLF terminators")
if line[0] in (SPACE, TAB):
# It's a continuation line.
v = line.strip()
else:
try:
k, v = line.split(COLON, 1)
except ValueError:
raise ValueError("Illegal header line.")
# TODO: what about TE and WWW-Authenticate?
k = k.strip().title()
v = v.strip()
hname = k
if k in comma_separated_headers:
existing = hdict.get(hname)
if existing:
v = ", ".join((existing, v))
hdict[hname] = v
return hdict
class MaxSizeExceeded(Exception):
pass
class SizeCheckWrapper(object):
"""Wraps a file-like object, raising MaxSizeExceeded if too large."""
def __init__(self, rfile, maxlen):
self.rfile = rfile
self.maxlen = maxlen
self.bytes_read = 0
def _check_length(self):
if self.maxlen and self.bytes_read > self.maxlen:
raise MaxSizeExceeded()
def read(self, size=None):
data = self.rfile.read(size)
self.bytes_read += len(data)
self._check_length()
return data
def readline(self, size=None):
if size is not None:
data = self.rfile.readline(size)
self.bytes_read += len(data)
self._check_length()
return data
# User didn't specify a size ...
# We read the line in chunks to make sure it's not a 100MB line !
res = []
while True:
data = self.rfile.readline(256)
self.bytes_read += len(data)
self._check_length()
res.append(data)
# See https://bitbucket.org/cherrypy/cherrypy/issue/421
if len(data) < 256 or data[-1:] == LF:
return EMPTY.join(res)
def readlines(self, sizehint=0):
# Shamelessly stolen from StringIO
total = 0
lines = []
line = self.readline()
while line:
lines.append(line)
total += len(line)
if 0 < sizehint <= total:
break
line = self.readline()
return lines
def close(self):
self.rfile.close()
def __iter__(self):
return self
def __next__(self):
data = next(self.rfile)
self.bytes_read += len(data)
self._check_length()
return data
def next(self):
data = self.rfile.next()
self.bytes_read += len(data)
self._check_length()
return data
class KnownLengthRFile(object):
"""Wraps a file-like object, returning an empty string when exhausted."""
def __init__(self, rfile, content_length):
self.rfile = rfile
self.remaining = content_length
def read(self, size=None):
if self.remaining == 0:
return ''
if size is None:
size = self.remaining
else:
size = min(size, self.remaining)
data = self.rfile.read(size)
self.remaining -= len(data)
return data
def readline(self, size=None):
if self.remaining == 0:
return ''
if size is None:
size = self.remaining
else:
size = min(size, self.remaining)
data = self.rfile.readline(size)
self.remaining -= len(data)
return data
def readlines(self, sizehint=0):
# Shamelessly stolen from StringIO
total = 0
lines = []
line = self.readline(sizehint)
while line:
lines.append(line)
total += len(line)
if 0 < sizehint <= total:
break
line = self.readline(sizehint)
return lines
def close(self):
self.rfile.close()
def __iter__(self):
return self
def __next__(self):
data = next(self.rfile)
self.remaining -= len(data)
return data
class ChunkedRFile(object):
"""Wraps a file-like object, returning an empty string when exhausted.
This class is intended to provide a conforming wsgi.input value for
request entities that have been encoded with the 'chunked' transfer
encoding.
"""
def __init__(self, rfile, maxlen, bufsize=8192):
self.rfile = rfile
self.maxlen = maxlen
self.bytes_read = 0
self.buffer = EMPTY
self.bufsize = bufsize
self.closed = False
def _fetch(self):
if self.closed:
return
line = self.rfile.readline()
self.bytes_read += len(line)
if self.maxlen and self.bytes_read > self.maxlen:
raise MaxSizeExceeded("Request Entity Too Large", self.maxlen)
line = line.strip().split(SEMICOLON, 1)
try:
chunk_size = line.pop(0)
chunk_size = int(chunk_size, 16)
except ValueError:
raise ValueError("Bad chunked transfer size: " + repr(chunk_size))
if chunk_size <= 0:
self.closed = True
return
## if line: chunk_extension = line[0]
if self.maxlen and self.bytes_read + chunk_size > self.maxlen:
raise IOError("Request Entity Too Large")
chunk = self.rfile.read(chunk_size)
self.bytes_read += len(chunk)
self.buffer += chunk
crlf = self.rfile.read(2)
if crlf != CRLF:
raise ValueError(
"Bad chunked transfer coding (expected '\\r\\n', "
"got " + repr(crlf) + ")")
def read(self, size=None):
data = EMPTY
while True:
if size and len(data) >= size:
return data
if not self.buffer:
self._fetch()
if not self.buffer:
# EOF
return data
if size:
remaining = size - len(data)
data += self.buffer[:remaining]
self.buffer = self.buffer[remaining:]
else:
data += self.buffer
def readline(self, size=None):
data = EMPTY
while True:
if size and len(data) >= size:
return data
if not self.buffer:
self._fetch()
if not self.buffer:
# EOF
return data
newline_pos = self.buffer.find(LF)
if size:
if newline_pos == -1:
remaining = size - len(data)
data += self.buffer[:remaining]
self.buffer = self.buffer[remaining:]
else:
remaining = min(size - len(data), newline_pos)
data += self.buffer[:remaining]
self.buffer = self.buffer[remaining:]
else:
if newline_pos == -1:
data += self.buffer
else:
data += self.buffer[:newline_pos]
self.buffer = self.buffer[newline_pos:]
def readlines(self, sizehint=0):
# Shamelessly stolen from StringIO
total = 0
lines = []
line = self.readline(sizehint)
while line:
lines.append(line)
total += len(line)
if 0 < sizehint <= total:
break
line = self.readline(sizehint)
return lines
def read_trailer_lines(self):
if not self.closed:
raise ValueError(
"Cannot read trailers until the request body has been read.")
while True:
line = self.rfile.readline()
if not line:
# No more data--illegal end of headers
raise ValueError("Illegal end of headers.")
self.bytes_read += len(line)
if self.maxlen and self.bytes_read > self.maxlen:
raise IOError("Request Entity Too Large")
if line == CRLF:
# Normal end of headers
break
if not line.endswith(CRLF):
raise ValueError("HTTP requires CRLF terminators")
yield line
def close(self):
self.rfile.close()
def __iter__(self):
# Shamelessly stolen from StringIO
total = 0
line = self.readline(sizehint)
while line:
yield line
total += len(line)
if 0 < sizehint <= total:
break
line = self.readline(sizehint)
class HTTPRequest(object):
"""An HTTP Request (and response).
A single HTTP connection may consist of multiple request/response pairs.
"""
server = None
"""The HTTPServer object which is receiving this request."""
conn = None
"""The HTTPConnection object on which this request connected."""
inheaders = {}
"""A dict of request headers."""
outheaders = []
"""A list of header tuples to write in the response."""
ready = False
"""When True, the request has been parsed and is ready to begin generating
the response. When False, signals the calling Connection that the response
should not be generated and the connection should close."""
close_connection = False
"""Signals the calling Connection that the request should close. This does
not imply an error! The client and/or server may each request that the
connection be closed."""
chunked_write = False
"""If True, output will be encoded with the "chunked" transfer-coding.
This value is set automatically inside send_headers."""
def __init__(self, server, conn):
self.server = server
self.conn = conn
self.ready = False
self.started_request = False
self.scheme = ntob("http")
if self.server.ssl_adapter is not None:
self.scheme = ntob("https")
# Use the lowest-common protocol in case read_request_line errors.
self.response_protocol = 'HTTP/1.0'
self.inheaders = {}
self.status = ""
self.outheaders = []
self.sent_headers = False
self.close_connection = self.__class__.close_connection
self.chunked_read = False
self.chunked_write = self.__class__.chunked_write
def parse_request(self):
"""Parse the next HTTP request start-line and message-headers."""
self.rfile = SizeCheckWrapper(self.conn.rfile,
self.server.max_request_header_size)
try:
success = self.read_request_line()
except MaxSizeExceeded:
self.simple_response(
"414 Request-URI Too Long",
"The Request-URI sent with the request exceeds the maximum "
"allowed bytes.")
return
else:
if not success:
return
try:
success = self.read_request_headers()
except MaxSizeExceeded:
self.simple_response(
"413 Request Entity Too Large",
"The headers sent with the request exceed the maximum "
"allowed bytes.")
return
else:
if not success:
return
self.ready = True
def read_request_line(self):
# HTTP/1.1 connections are persistent by default. If a client
# requests a page, then idles (leaves the connection open),
# then rfile.readline() will raise socket.error("timed out").
# Note that it does this based on the value given to settimeout(),
# and doesn't need the client to request or acknowledge the close
# (although your TCP stack might suffer for it: cf Apache's history
# with FIN_WAIT_2).
request_line = self.rfile.readline()
# Set started_request to True so communicate() knows to send 408
# from here on out.
self.started_request = True
if not request_line:
return False
if request_line == CRLF:
# RFC 2616 sec 4.1: "...if the server is reading the protocol
# stream at the beginning of a message and receives a CRLF
# first, it should ignore the CRLF."
# But only ignore one leading line! else we enable a DoS.
request_line = self.rfile.readline()
if not request_line:
return False
if not request_line.endswith(CRLF):
self.simple_response(
"400 Bad Request", "HTTP requires CRLF terminators")
return False
try:
method, uri, req_protocol = request_line.strip().split(SPACE, 2)
rp = int(req_protocol[5]), int(req_protocol[7])
except (ValueError, IndexError):
self.simple_response("400 Bad Request", "Malformed Request-Line")
return False
self.uri = uri
self.method = method
# uri may be an abs_path (including "http://host.domain.tld");
scheme, authority, path = self.parse_request_uri(uri)
if NUMBER_SIGN in path:
self.simple_response("400 Bad Request",
"Illegal #fragment in Request-URI.")
return False
if scheme:
self.scheme = scheme
qs = EMPTY
if QUESTION_MARK in path:
path, qs = path.split(QUESTION_MARK, 1)
# Unquote the path+params (e.g. "/this%20path" -> "/this path").
# http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2
#
# But note that "...a URI must be separated into its components
# before the escaped characters within those components can be
# safely decoded." http://www.ietf.org/rfc/rfc2396.txt, sec 2.4.2
# Therefore, "/this%2Fpath" becomes "/this%2Fpath", not "/this/path".
try:
atoms = [unquote(x) for x in quoted_slash.split(path)]
except ValueError:
ex = sys.exc_info()[1]
self.simple_response("400 Bad Request", ex.args[0])
return False
path = "%2F".join(atoms)
self.path = path
# Note that, like wsgiref and most other HTTP servers,
# we "% HEX HEX"-unquote the path but not the query string.
self.qs = qs
# Compare request and server HTTP protocol versions, in case our
# server does not support the requested protocol. Limit our output
# to min(req, server). We want the following output:
# request server actual written supported response
# protocol protocol response protocol feature set
# a 1.0 1.0 1.0 1.0
# b 1.0 1.1 1.1 1.0
# c 1.1 1.0 1.0 1.0
# d 1.1 1.1 1.1 1.1
# Notice that, in (b), the response will be "HTTP/1.1" even though
# the client only understands 1.0. RFC 2616 10.5.6 says we should
# only return 505 if the _major_ version is different.
sp = int(self.server.protocol[5]), int(self.server.protocol[7])
if sp[0] != rp[0]:
self.simple_response("505 HTTP Version Not Supported")
return False
self.request_protocol = req_protocol
self.response_protocol = "HTTP/%s.%s" % min(rp, sp)
return True
def read_request_headers(self):
"""Read self.rfile into self.inheaders. Return success."""
# then all the http headers
try:
read_headers(self.rfile, self.inheaders)
except ValueError:
ex = sys.exc_info()[1]
self.simple_response("400 Bad Request", ex.args[0])
return False
mrbs = self.server.max_request_body_size
if mrbs and int(self.inheaders.get("Content-Length", 0)) > mrbs:
self.simple_response(
"413 Request Entity Too Large",
"The entity sent with the request exceeds the maximum "
"allowed bytes.")
return False
# Persistent connection support
if self.response_protocol == "HTTP/1.1":
# Both server and client are HTTP/1.1
if self.inheaders.get("Connection", "") == "close":
self.close_connection = True
else:
# Either the server or client (or both) are HTTP/1.0
if self.inheaders.get("Connection", "") != "Keep-Alive":
self.close_connection = True
# Transfer-Encoding support
te = None
if self.response_protocol == "HTTP/1.1":
te = self.inheaders.get("Transfer-Encoding")
if te:
te = [x.strip().lower() for x in te.split(",") if x.strip()]
self.chunked_read = False
if te:
for enc in te:
if enc == "chunked":
self.chunked_read = True
else:
# Note that, even if we see "chunked", we must reject
# if there is an extension we don't recognize.
self.simple_response("501 Unimplemented")
self.close_connection = True
return False
# From PEP 333:
# "Servers and gateways that implement HTTP 1.1 must provide
# transparent support for HTTP 1.1's "expect/continue" mechanism.
# This may be done in any of several ways:
# 1. Respond to requests containing an Expect: 100-continue request
# with an immediate "100 Continue" response, and proceed normally.
# 2. Proceed with the request normally, but provide the application
# with a wsgi.input stream that will send the "100 Continue"
# response if/when the application first attempts to read from
# the input stream. The read request must then remain blocked
# until the client responds.
# 3. Wait until the client decides that the server does not support
# expect/continue, and sends the request body on its own.
# (This is suboptimal, and is not recommended.)
#
# We used to do 3, but are now doing 1. Maybe we'll do 2 someday,
# but it seems like it would be a big slowdown for such a rare case.
if self.inheaders.get("Expect", "") == "100-continue":
# Don't use simple_response here, because it emits headers
# we don't want. See
# https://bitbucket.org/cherrypy/cherrypy/issue/951
msg = self.server.protocol + " 100 Continue\r\n\r\n"
try:
self.conn.wfile.sendall(msg)
except socket.error:
x = sys.exc_info()[1]
if x.args[0] not in socket_errors_to_ignore:
raise
return True
def parse_request_uri(self, uri):
"""Parse a Request-URI into (scheme, authority, path).
Note that Request-URI's must be one of::
Request-URI = "*" | absoluteURI | abs_path | authority
Therefore, a Request-URI which starts with a double forward-slash
cannot be a "net_path"::
net_path = "//" authority [ abs_path ]
Instead, it must be interpreted as an "abs_path" with an empty first
path segment::
abs_path = "/" path_segments
path_segments = segment *( "/" segment )
segment = *pchar *( ";" param )
param = *pchar
"""
if uri == ASTERISK:
return None, None, uri
i = uri.find('://')
if i > 0 and QUESTION_MARK not in uri[:i]:
# An absoluteURI.
# If there's a scheme (and it must be http or https), then:
# http_URL = "http:" "//" host [ ":" port ] [ abs_path [ "?" query
# ]]
scheme, remainder = uri[:i].lower(), uri[i + 3:]
authority, path = remainder.split(FORWARD_SLASH, 1)
path = FORWARD_SLASH + path
return scheme, authority, path
if uri.startswith(FORWARD_SLASH):
# An abs_path.
return None, None, uri
else:
# An authority.
return None, uri, None
def respond(self):
"""Call the gateway and write its iterable output."""
mrbs = self.server.max_request_body_size
if self.chunked_read:
self.rfile = ChunkedRFile(self.conn.rfile, mrbs)
else:
cl = int(self.inheaders.get("Content-Length", 0))
if mrbs and mrbs < cl:
if not self.sent_headers:
self.simple_response(
"413 Request Entity Too Large",
"The entity sent with the request exceeds the maximum "
"allowed bytes.")
return
self.rfile = KnownLengthRFile(self.conn.rfile, cl)
self.server.gateway(self).respond()
if (self.ready and not self.sent_headers):
self.sent_headers = True
self.send_headers()
if self.chunked_write:
self.conn.wfile.sendall("0\r\n\r\n")
def simple_response(self, status, msg=""):
"""Write a simple response back to the client."""
status = str(status)
buf = [self.server.protocol + SPACE +
status + CRLF,
"Content-Length: %s\r\n" % len(msg),
"Content-Type: text/plain\r\n"]
if status[:3] in ("413", "414"):
# Request Entity Too Large / Request-URI Too Long
self.close_connection = True
if self.response_protocol == 'HTTP/1.1':
# This will not be true for 414, since read_request_line
# usually raises 414 before reading the whole line, and we
# therefore cannot know the proper response_protocol.
buf.append("Connection: close\r\n")
else:
# HTTP/1.0 had no 413/414 status nor Connection header.
# Emit 400 instead and trust the message body is enough.
status = "400 Bad Request"
buf.append(CRLF)
if msg:
if isinstance(msg, unicodestr):
msg = msg.encode("ISO-8859-1")
buf.append(msg)
try:
self.conn.wfile.sendall("".join(buf))
except socket.error:
x = sys.exc_info()[1]
if x.args[0] not in socket_errors_to_ignore:
raise
def write(self, chunk):
"""Write unbuffered data to the client."""
if self.chunked_write and chunk:
buf = [hex(len(chunk))[2:], CRLF, chunk, CRLF]
self.conn.wfile.sendall(EMPTY.join(buf))
else:
self.conn.wfile.sendall(chunk)
def send_headers(self):
"""Assert, process, and send the HTTP response message-headers.
You must set self.status, and self.outheaders before calling this.
"""
hkeys = [key.lower() for key, value in self.outheaders]
status = int(self.status[:3])
if status == 413:
# Request Entity Too Large. Close conn to avoid garbage.
self.close_connection = True
elif "content-length" not in hkeys:
# "All 1xx (informational), 204 (no content),
# and 304 (not modified) responses MUST NOT
# include a message-body." So no point chunking.
if status < 200 or status in (204, 205, 304):
pass
else:
if (self.response_protocol == 'HTTP/1.1'
and self.method != 'HEAD'):
# Use the chunked transfer-coding
self.chunked_write = True
self.outheaders.append(("Transfer-Encoding", "chunked"))
else:
# Closing the conn is the only way to determine len.
self.close_connection = True
if "connection" not in hkeys:
if self.response_protocol == 'HTTP/1.1':
# Both server and client are HTTP/1.1 or better
if self.close_connection:
self.outheaders.append(("Connection", "close"))
else:
# Server and/or client are HTTP/1.0
if not self.close_connection:
self.outheaders.append(("Connection", "Keep-Alive"))
if (not self.close_connection) and (not self.chunked_read):
# Read any remaining request body data on the socket.
# "If an origin server receives a request that does not include an
# Expect request-header field with the "100-continue" expectation,
# the request includes a request body, and the server responds
# with a final status code before reading the entire request body
# from the transport connection, then the server SHOULD NOT close
# the transport connection until it has read the entire request,
# or until the client closes the connection. Otherwise, the client
# might not reliably receive the response message. However, this
# requirement is not be construed as preventing a server from
# defending itself against denial-of-service attacks, or from
# badly broken client implementations."
remaining = getattr(self.rfile, 'remaining', 0)
if remaining > 0:
self.rfile.read(remaining)
if "date" not in hkeys:
self.outheaders.append(("Date", rfc822.formatdate()))
if "server" not in hkeys:
self.outheaders.append(("Server", self.server.server_name))
buf = [self.server.protocol + SPACE + self.status + CRLF]
for k, v in self.outheaders:
buf.append(k + COLON + SPACE + v + CRLF)
buf.append(CRLF)
self.conn.wfile.sendall(EMPTY.join(buf))
class NoSSLError(Exception):
"""Exception raised when a client speaks HTTP to an HTTPS socket."""
pass
class FatalSSLAlert(Exception):
"""Exception raised when the SSL implementation signals a fatal alert."""
pass
class CP_fileobject(socket._fileobject):
"""Faux file object attached to a socket object."""
def __init__(self, *args, **kwargs):
self.bytes_read = 0
self.bytes_written = 0
socket._fileobject.__init__(self, *args, **kwargs)
def sendall(self, data):
"""Sendall for non-blocking sockets."""
while data:
try:
bytes_sent = self.send(data)
data = data[bytes_sent:]
except socket.error, e:
if e.args[0] not in socket_errors_nonblocking:
raise
def send(self, data):
bytes_sent = self._sock.send(data)
self.bytes_written += bytes_sent
return bytes_sent
def flush(self):
if self._wbuf:
buffer = "".join(self._wbuf)
self._wbuf = []
self.sendall(buffer)
def recv(self, size):
while True:
try:
data = self._sock.recv(size)
self.bytes_read += len(data)
return data
except socket.error, e:
if (e.args[0] not in socket_errors_nonblocking
and e.args[0] not in socket_error_eintr):
raise
if not _fileobject_uses_str_type:
def read(self, size=-1):
# Use max, disallow tiny reads in a loop as they are very
# inefficient.
# We never leave read() with any leftover data from a new recv()
# call in our internal buffer.
rbufsize = max(self._rbufsize, self.default_bufsize)
# Our use of StringIO rather than lists of string objects returned
# by recv() minimizes memory usage and fragmentation that occurs
# when rbufsize is large compared to the typical return value of
# recv().
buf = self._rbuf
buf.seek(0, 2) # seek end
if size < 0:
# Read until EOF
# reset _rbuf. we consume it via buf.
self._rbuf = StringIO.StringIO()
while True:
data = self.recv(rbufsize)
if not data:
break
buf.write(data)
return buf.getvalue()
else:
# Read until size bytes or EOF seen, whichever comes first
buf_len = buf.tell()
if buf_len >= size:
# Already have size bytes in our buffer? Extract and
# return.
buf.seek(0)
rv = buf.read(size)
self._rbuf = StringIO.StringIO()
self._rbuf.write(buf.read())
return rv
# reset _rbuf. we consume it via buf.
self._rbuf = StringIO.StringIO()
while True:
left = size - buf_len
# recv() will malloc the amount of memory given as its
# parameter even though it often returns much less data
# than that. The returned data string is short lived
# as we copy it into a StringIO and free it. This avoids
# fragmentation issues on many platforms.
data = self.recv(left)
if not data:
break
n = len(data)
if n == size and not buf_len:
# Shortcut. Avoid buffer data copies when:
# - We have no data in our buffer.
# AND
# - Our call to recv returned exactly the
# number of bytes we were asked to read.
return data
if n == left:
buf.write(data)
del data # explicit free
break
assert n <= left, "recv(%d) returned %d bytes" % (left, n)
buf.write(data)
buf_len += n
del data # explicit free
#assert buf_len == buf.tell()
return buf.getvalue()
def readline(self, size=-1):
buf = self._rbuf
buf.seek(0, 2) # seek end
if buf.tell() > 0:
# check if we already have it in our buffer
buf.seek(0)
bline = buf.readline(size)
if bline.endswith('\n') or len(bline) == size:
self._rbuf = StringIO.StringIO()
self._rbuf.write(buf.read())
return bline
del bline
if size < 0:
# Read until \n or EOF, whichever comes first
if self._rbufsize <= 1:
# Speed up unbuffered case
buf.seek(0)
buffers = [buf.read()]
# reset _rbuf. we consume it via buf.
self._rbuf = StringIO.StringIO()
data = None
recv = self.recv
while data != "\n":
data = recv(1)
if not data:
break
buffers.append(data)
return "".join(buffers)
buf.seek(0, 2) # seek end
# reset _rbuf. we consume it via buf.
self._rbuf = StringIO.StringIO()
while True:
data = self.recv(self._rbufsize)
if not data:
break
nl = data.find('\n')
if nl >= 0:
nl += 1
buf.write(data[:nl])
self._rbuf.write(data[nl:])
del data
break
buf.write(data)
return buf.getvalue()
else:
# Read until size bytes or \n or EOF seen, whichever comes
# first
buf.seek(0, 2) # seek end
buf_len = buf.tell()
if buf_len >= size:
buf.seek(0)
rv = buf.read(size)
self._rbuf = StringIO.StringIO()
self._rbuf.write(buf.read())
return rv
# reset _rbuf. we consume it via buf.
self._rbuf = StringIO.StringIO()
while True:
data = self.recv(self._rbufsize)
if not data:
break
left = size - buf_len
# did we just receive a newline?
nl = data.find('\n', 0, left)
if nl >= 0:
nl += 1
# save the excess data to _rbuf
self._rbuf.write(data[nl:])
if buf_len:
buf.write(data[:nl])
break
else:
# Shortcut. Avoid data copy through buf when
# returning a substring of our first recv().
return data[:nl]
n = len(data)
if n == size and not buf_len:
# Shortcut. Avoid data copy through buf when
# returning exactly all of our first recv().
return data
if n >= left:
buf.write(data[:left])
self._rbuf.write(data[left:])
break
buf.write(data)
buf_len += n
#assert buf_len == buf.tell()
return buf.getvalue()
else:
def read(self, size=-1):
if size < 0:
# Read until EOF
buffers = [self._rbuf]
self._rbuf = ""
if self._rbufsize <= 1:
recv_size = self.default_bufsize
else:
recv_size = self._rbufsize
while True:
data = self.recv(recv_size)
if not data:
break
buffers.append(data)
return "".join(buffers)
else:
# Read until size bytes or EOF seen, whichever comes first
data = self._rbuf
buf_len = len(data)
if buf_len >= size:
self._rbuf = data[size:]
return data[:size]
buffers = []
if data:
buffers.append(data)
self._rbuf = ""
while True:
left = size - buf_len
recv_size = max(self._rbufsize, left)
data = self.recv(recv_size)
if not data:
break
buffers.append(data)
n = len(data)
if n >= left:
self._rbuf = data[left:]
buffers[-1] = data[:left]
break
buf_len += n
return "".join(buffers)
def readline(self, size=-1):
data = self._rbuf
if size < 0:
# Read until \n or EOF, whichever comes first
if self._rbufsize <= 1:
# Speed up unbuffered case
assert data == ""
buffers = []
while data != "\n":
data = self.recv(1)
if not data:
break
buffers.append(data)
return "".join(buffers)
nl = data.find('\n')
if nl >= 0:
nl += 1
self._rbuf = data[nl:]
return data[:nl]
buffers = []
if data:
buffers.append(data)
self._rbuf = ""
while True:
data = self.recv(self._rbufsize)
if not data:
break
buffers.append(data)
nl = data.find('\n')
if nl >= 0:
nl += 1
self._rbuf = data[nl:]
buffers[-1] = data[:nl]
break
return "".join(buffers)
else:
# Read until size bytes or \n or EOF seen, whichever comes
# first
nl = data.find('\n', 0, size)
if nl >= 0:
nl += 1
self._rbuf = data[nl:]
return data[:nl]
buf_len = len(data)
if buf_len >= size:
self._rbuf = data[size:]
return data[:size]
buffers = []
if data:
buffers.append(data)
self._rbuf = ""
while True:
data = self.recv(self._rbufsize)
if not data:
break
buffers.append(data)
left = size - buf_len
nl = data.find('\n', 0, left)
if nl >= 0:
nl += 1
self._rbuf = data[nl:]
buffers[-1] = data[:nl]
break
n = len(data)
if n >= left:
self._rbuf = data[left:]
buffers[-1] = data[:left]
break
buf_len += n
return "".join(buffers)
class HTTPConnection(object):
"""An HTTP connection (active socket).
server: the Server object which received this connection.
socket: the raw socket object (usually TCP) for this connection.
makefile: a fileobject class for reading from the socket.
"""
remote_addr = None
remote_port = None
ssl_env = None
rbufsize = DEFAULT_BUFFER_SIZE
wbufsize = DEFAULT_BUFFER_SIZE
RequestHandlerClass = HTTPRequest
def __init__(self, server, sock, makefile=CP_fileobject):
self.server = server
self.socket = sock
self.rfile = makefile(sock, "rb", self.rbufsize)
self.wfile = makefile(sock, "wb", self.wbufsize)
self.requests_seen = 0
def communicate(self):
"""Read each request and respond appropriately."""
request_seen = False
try:
while True:
# (re)set req to None so that if something goes wrong in
# the RequestHandlerClass constructor, the error doesn't
# get written to the previous request.
req = None
req = self.RequestHandlerClass(self.server, self)
# This order of operations should guarantee correct pipelining.
req.parse_request()
if self.server.stats['Enabled']:
self.requests_seen += 1
if not req.ready:
# Something went wrong in the parsing (and the server has
# probably already made a simple_response). Return and
# let the conn close.
return
request_seen = True
req.respond()
if req.close_connection:
return
except socket.error:
e = sys.exc_info()[1]
errnum = e.args[0]
# sadly SSL sockets return a different (longer) time out string
if (
errnum == 'timed out' or
errnum == 'The read operation timed out'
):
# Don't error if we're between requests; only error
# if 1) no request has been started at all, or 2) we're
# in the middle of a request.
# See https://bitbucket.org/cherrypy/cherrypy/issue/853
if (not request_seen) or (req and req.started_request):
# Don't bother writing the 408 if the response
# has already started being written.
if req and not req.sent_headers:
try:
req.simple_response("408 Request Timeout")
except FatalSSLAlert:
# Close the connection.
return
elif errnum not in socket_errors_to_ignore:
self.server.error_log("socket.error %s" % repr(errnum),
level=logging.WARNING, traceback=True)
if req and not req.sent_headers:
try:
req.simple_response("500 Internal Server Error")
except FatalSSLAlert:
# Close the connection.
return
return
except (KeyboardInterrupt, SystemExit):
raise
except FatalSSLAlert:
# Close the connection.
return
except NoSSLError:
if req and not req.sent_headers:
# Unwrap our wfile
self.wfile = CP_fileobject(
self.socket._sock, "wb", self.wbufsize)
req.simple_response(
"400 Bad Request",
"The client sent a plain HTTP request, but "
"this server only speaks HTTPS on this port.")
self.linger = True
except Exception:
e = sys.exc_info()[1]
self.server.error_log(repr(e), level=logging.ERROR, traceback=True)
if req and not req.sent_headers:
try:
req.simple_response("500 Internal Server Error")
except FatalSSLAlert:
# Close the connection.
return
linger = False
def close(self):
"""Close the socket underlying this connection."""
self.rfile.close()
if not self.linger:
# Python's socket module does NOT call close on the kernel
# socket when you call socket.close(). We do so manually here
# because we want this server to send a FIN TCP segment
# immediately. Note this must be called *before* calling
# socket.close(), because the latter drops its reference to
# the kernel socket.
if hasattr(self.socket, '_sock'):
self.socket._sock.close()
self.socket.close()
else:
# On the other hand, sometimes we want to hang around for a bit
# to make sure the client has a chance to read our entire
# response. Skipping the close() calls here delays the FIN
# packet until the socket object is garbage-collected later.
# Someday, perhaps, we'll do the full lingering_close that
# Apache does, but not today.
pass
class TrueyZero(object):
"""An object which equals and does math like the integer 0 but evals True.
"""
def __add__(self, other):
return other
def __radd__(self, other):
return other
trueyzero = TrueyZero()
_SHUTDOWNREQUEST = None
class WorkerThread(threading.Thread):
"""Thread which continuously polls a Queue for Connection objects.
Due to the timing issues of polling a Queue, a WorkerThread does not
check its own 'ready' flag after it has started. To stop the thread,
it is necessary to stick a _SHUTDOWNREQUEST object onto the Queue
(one for each running WorkerThread).
"""
conn = None
"""The current connection pulled off the Queue, or None."""
server = None
"""The HTTP Server which spawned this thread, and which owns the
Queue and is placing active connections into it."""
ready = False
"""A simple flag for the calling server to know when this thread
has begun polling the Queue."""
def __init__(self, server):
self.ready = False
self.server = server
self.requests_seen = 0
self.bytes_read = 0
self.bytes_written = 0
self.start_time = None
self.work_time = 0
self.stats = {
'Requests': lambda s: self.requests_seen + (
(self.start_time is None) and
trueyzero or
self.conn.requests_seen
),
'Bytes Read': lambda s: self.bytes_read + (
(self.start_time is None) and
trueyzero or
self.conn.rfile.bytes_read
),
'Bytes Written': lambda s: self.bytes_written + (
(self.start_time is None) and
trueyzero or
self.conn.wfile.bytes_written
),
'Work Time': lambda s: self.work_time + (
(self.start_time is None) and
trueyzero or
time.time() - self.start_time
),
'Read Throughput': lambda s: s['Bytes Read'](s) / (
s['Work Time'](s) or 1e-6),
'Write Throughput': lambda s: s['Bytes Written'](s) / (
s['Work Time'](s) or 1e-6),
}
threading.Thread.__init__(self)
def run(self):
self.server.stats['Worker Threads'][self.getName()] = self.stats
try:
self.ready = True
while True:
conn = self.server.requests.get()
if conn is _SHUTDOWNREQUEST:
return
self.conn = conn
if self.server.stats['Enabled']:
self.start_time = time.time()
try:
conn.communicate()
finally:
conn.close()
if self.server.stats['Enabled']:
self.requests_seen += self.conn.requests_seen
self.bytes_read += self.conn.rfile.bytes_read
self.bytes_written += self.conn.wfile.bytes_written
self.work_time += time.time() - self.start_time
self.start_time = None
self.conn = None
except (KeyboardInterrupt, SystemExit):
exc = sys.exc_info()[1]
self.server.interrupt = exc
class ThreadPool(object):
"""A Request Queue for an HTTPServer which pools threads.
ThreadPool objects must provide min, get(), put(obj), start()
and stop(timeout) attributes.
"""
def __init__(self, server, min=10, max=-1,
accepted_queue_size=-1, accepted_queue_timeout=10):
self.server = server
self.min = min
self.max = max
self._threads = []
self._queue = queue.Queue(maxsize=accepted_queue_size)
self._queue_put_timeout = accepted_queue_timeout
self.get = self._queue.get
def start(self):
"""Start the pool of threads."""
for i in range(self.min):
self._threads.append(WorkerThread(self.server))
for worker in self._threads:
worker.setName("CP Server " + worker.getName())
worker.start()
for worker in self._threads:
while not worker.ready:
time.sleep(.1)
def _get_idle(self):
"""Number of worker threads which are idle. Read-only."""
return len([t for t in self._threads if t.conn is None])
idle = property(_get_idle, doc=_get_idle.__doc__)
def put(self, obj):
self._queue.put(obj, block=True, timeout=self._queue_put_timeout)
if obj is _SHUTDOWNREQUEST:
return
def grow(self, amount):
"""Spawn new worker threads (not above self.max)."""
if self.max > 0:
budget = max(self.max - len(self._threads), 0)
else:
# self.max <= 0 indicates no maximum
budget = float('inf')
n_new = min(amount, budget)
workers = [self._spawn_worker() for i in range(n_new)]
while not self._all(operator.attrgetter('ready'), workers):
time.sleep(.1)
self._threads.extend(workers)
def _spawn_worker(self):
worker = WorkerThread(self.server)
worker.setName("CP Server " + worker.getName())
worker.start()
return worker
def _all(func, items):
results = [func(item) for item in items]
return reduce(operator.and_, results, True)
_all = staticmethod(_all)
def shrink(self, amount):
"""Kill off worker threads (not below self.min)."""
# Grow/shrink the pool if necessary.
# Remove any dead threads from our list
for t in self._threads:
if not t.isAlive():
self._threads.remove(t)
amount -= 1
# calculate the number of threads above the minimum
n_extra = max(len(self._threads) - self.min, 0)
# don't remove more than amount
n_to_remove = min(amount, n_extra)
# put shutdown requests on the queue equal to the number of threads
# to remove. As each request is processed by a worker, that worker
# will terminate and be culled from the list.
for n in range(n_to_remove):
self._queue.put(_SHUTDOWNREQUEST)
def stop(self, timeout=5):
# Must shut down threads here so the code that calls
# this method can know when all threads are stopped.
for worker in self._threads:
self._queue.put(_SHUTDOWNREQUEST)
# Don't join currentThread (when stop is called inside a request).
current = threading.currentThread()
if timeout and timeout >= 0:
endtime = time.time() + timeout
while self._threads:
worker = self._threads.pop()
if worker is not current and worker.isAlive():
try:
if timeout is None or timeout < 0:
worker.join()
else:
remaining_time = endtime - time.time()
if remaining_time > 0:
worker.join(remaining_time)
if worker.isAlive():
# We exhausted the timeout.
# Forcibly shut down the socket.
c = worker.conn
if c and not c.rfile.closed:
try:
c.socket.shutdown(socket.SHUT_RD)
except TypeError:
# pyOpenSSL sockets don't take an arg
c.socket.shutdown()
worker.join()
except (AssertionError,
# Ignore repeated Ctrl-C.
# See
# https://bitbucket.org/cherrypy/cherrypy/issue/691.
KeyboardInterrupt):
pass
def _get_qsize(self):
return self._queue.qsize()
qsize = property(_get_qsize)
try:
import fcntl
except ImportError:
try:
import ctypes.wintypes
from ctypes import WinError, windll
_SetHandleInformation = windll.kernel32.SetHandleInformation
_SetHandleInformation.argtypes = [
ctypes.wintypes.HANDLE,
ctypes.wintypes.DWORD,
ctypes.wintypes.DWORD,
]
_SetHandleInformation.restype = ctypes.wintypes.BOOL
except ImportError:
def prevent_socket_inheritance(sock):
"""Dummy function, since neither fcntl nor ctypes are available."""
pass
else:
def prevent_socket_inheritance(sock):
"""Mark the given socket fd as non-inheritable (Windows)."""
if not _SetHandleInformation(sock.fileno(), 1, 0):
raise WinError()
else:
def prevent_socket_inheritance(sock):
"""Mark the given socket fd as non-inheritable (POSIX)."""
fd = sock.fileno()
old_flags = fcntl.fcntl(fd, fcntl.F_GETFD)
fcntl.fcntl(fd, fcntl.F_SETFD, old_flags | fcntl.FD_CLOEXEC)
class SSLAdapter(object):
"""Base class for SSL driver library adapters.
Required methods:
* ``wrap(sock) -> (wrapped socket, ssl environ dict)``
* ``makefile(sock, mode='r', bufsize=DEFAULT_BUFFER_SIZE) ->
socket file object``
"""
def __init__(self, certificate, private_key, certificate_chain=None):
self.certificate = certificate
self.private_key = private_key
self.certificate_chain = certificate_chain
def wrap(self, sock):
raise NotImplemented
def makefile(self, sock, mode='r', bufsize=DEFAULT_BUFFER_SIZE):
raise NotImplemented
class HTTPServer(object):
"""An HTTP server."""
_bind_addr = "127.0.0.1"
_interrupt = None
gateway = None
"""A Gateway instance."""
minthreads = None
"""The minimum number of worker threads to create (default 10)."""
maxthreads = None
"""The maximum number of worker threads to create (default -1 = no limit).
"""
server_name = None
"""The name of the server; defaults to socket.gethostname()."""
protocol = "HTTP/1.1"
"""The version string to write in the Status-Line of all HTTP responses.
For example, "HTTP/1.1" is the default. This also limits the supported
features used in the response."""
request_queue_size = 5
"""The 'backlog' arg to socket.listen(); max queued connections
(default 5).
"""
shutdown_timeout = 5
"""The total time, in seconds, to wait for worker threads to cleanly exit.
"""
timeout = 10
"""The timeout in seconds for accepted connections (default 10)."""
version = "CherryPy/3.6.0"
"""A version string for the HTTPServer."""
software = None
"""The value to set for the SERVER_SOFTWARE entry in the WSGI environ.
If None, this defaults to ``'%s Server' % self.version``."""
ready = False
"""An internal flag which marks whether the socket is accepting connections
"""
max_request_header_size = 0
"""The maximum size, in bytes, for request headers, or 0 for no limit."""
max_request_body_size = 0
"""The maximum size, in bytes, for request bodies, or 0 for no limit."""
nodelay = True
"""If True (the default since 3.1), sets the TCP_NODELAY socket option."""
ConnectionClass = HTTPConnection
"""The class to use for handling HTTP connections."""
ssl_adapter = None
"""An instance of SSLAdapter (or a subclass).
You must have the corresponding SSL driver library installed."""
def __init__(self, bind_addr, gateway, minthreads=10, maxthreads=-1,
server_name=None):
self.bind_addr = bind_addr
self.gateway = gateway
self.requests = ThreadPool(self, min=minthreads or 1, max=maxthreads)
if not server_name:
server_name = socket.gethostname()
self.server_name = server_name
self.clear_stats()
def clear_stats(self):
self._start_time = None
self._run_time = 0
self.stats = {
'Enabled': False,
'Bind Address': lambda s: repr(self.bind_addr),
'Run time': lambda s: (not s['Enabled']) and -1 or self.runtime(),
'Accepts': 0,
'Accepts/sec': lambda s: s['Accepts'] / self.runtime(),
'Queue': lambda s: getattr(self.requests, "qsize", None),
'Threads': lambda s: len(getattr(self.requests, "_threads", [])),
'Threads Idle': lambda s: getattr(self.requests, "idle", None),
'Socket Errors': 0,
'Requests': lambda s: (not s['Enabled']) and -1 or sum(
[w['Requests'](w) for w in s['Worker Threads'].values()], 0),
'Bytes Read': lambda s: (not s['Enabled']) and -1 or sum(
[w['Bytes Read'](w) for w in s['Worker Threads'].values()], 0),
'Bytes Written': lambda s: (not s['Enabled']) and -1 or sum(
[w['Bytes Written'](w) for w in s['Worker Threads'].values()],
0),
'Work Time': lambda s: (not s['Enabled']) and -1 or sum(
[w['Work Time'](w) for w in s['Worker Threads'].values()], 0),
'Read Throughput': lambda s: (not s['Enabled']) and -1 or sum(
[w['Bytes Read'](w) / (w['Work Time'](w) or 1e-6)
for w in s['Worker Threads'].values()], 0),
'Write Throughput': lambda s: (not s['Enabled']) and -1 or sum(
[w['Bytes Written'](w) / (w['Work Time'](w) or 1e-6)
for w in s['Worker Threads'].values()], 0),
'Worker Threads': {},
}
logging.statistics["CherryPy HTTPServer %d" % id(self)] = self.stats
def runtime(self):
if self._start_time is None:
return self._run_time
else:
return self._run_time + (time.time() - self._start_time)
def __str__(self):
return "%s.%s(%r)" % (self.__module__, self.__class__.__name__,
self.bind_addr)
def _get_bind_addr(self):
return self._bind_addr
def _set_bind_addr(self, value):
if isinstance(value, tuple) and value[0] in ('', None):
# Despite the socket module docs, using '' does not
# allow AI_PASSIVE to work. Passing None instead
# returns '0.0.0.0' like we want. In other words:
# host AI_PASSIVE result
# '' Y 192.168.x.y
# '' N 192.168.x.y
# None Y 0.0.0.0
# None N 127.0.0.1
# But since you can get the same effect with an explicit
# '0.0.0.0', we deny both the empty string and None as values.
raise ValueError("Host values of '' or None are not allowed. "
"Use '0.0.0.0' (IPv4) or '::' (IPv6) instead "
"to listen on all active interfaces.")
self._bind_addr = value
bind_addr = property(
_get_bind_addr,
_set_bind_addr,
doc="""The interface on which to listen for connections.
For TCP sockets, a (host, port) tuple. Host values may be any IPv4
or IPv6 address, or any valid hostname. The string 'localhost' is a
synonym for '127.0.0.1' (or '::1', if your hosts file prefers IPv6).
The string '0.0.0.0' is a special IPv4 entry meaning "any active
interface" (INADDR_ANY), and '::' is the similar IN6ADDR_ANY for
IPv6. The empty string or None are not allowed.
For UNIX sockets, supply the filename as a string.""")
def start(self):
"""Run the server forever."""
# We don't have to trap KeyboardInterrupt or SystemExit here,
# because cherrpy.server already does so, calling self.stop() for us.
# If you're using this server with another framework, you should
# trap those exceptions in whatever code block calls start().
self._interrupt = None
if self.software is None:
self.software = "%s Server" % self.version
# SSL backward compatibility
if (self.ssl_adapter is None and
getattr(self, 'ssl_certificate', None) and
getattr(self, 'ssl_private_key', None)):
warnings.warn(
"SSL attributes are deprecated in CherryPy 3.2, and will "
"be removed in CherryPy 3.3. Use an ssl_adapter attribute "
"instead.",
DeprecationWarning
)
try:
from cherrypy.wsgiserver.ssl_pyopenssl import pyOpenSSLAdapter
except ImportError:
pass
else:
self.ssl_adapter = pyOpenSSLAdapter(
self.ssl_certificate, self.ssl_private_key,
getattr(self, 'ssl_certificate_chain', None))
# Select the appropriate socket
if isinstance(self.bind_addr, basestring):
# AF_UNIX socket
# So we can reuse the socket...
try:
os.unlink(self.bind_addr)
except:
pass
# So everyone can access the socket...
try:
os.chmod(self.bind_addr, 511) # 0777
except:
pass
info = [
(socket.AF_UNIX, socket.SOCK_STREAM, 0, "", self.bind_addr)]
else:
# AF_INET or AF_INET6 socket
# Get the correct address family for our host (allows IPv6
# addresses)
host, port = self.bind_addr
try:
info = socket.getaddrinfo(
host, port, socket.AF_UNSPEC,
socket.SOCK_STREAM, 0, socket.AI_PASSIVE)
except socket.gaierror:
if ':' in self.bind_addr[0]:
info = [(socket.AF_INET6, socket.SOCK_STREAM,
0, "", self.bind_addr + (0, 0))]
else:
info = [(socket.AF_INET, socket.SOCK_STREAM,
0, "", self.bind_addr)]
self.socket = None
msg = "No socket could be created"
for res in info:
af, socktype, proto, canonname, sa = res
try:
self.bind(af, socktype, proto)
except socket.error, serr:
msg = "%s -- (%s: %s)" % (msg, sa, serr)
if self.socket:
self.socket.close()
self.socket = None
continue
break
if not self.socket:
raise socket.error(msg)
# Timeout so KeyboardInterrupt can be caught on Win32
self.socket.settimeout(1)
self.socket.listen(self.request_queue_size)
# Create worker threads
self.requests.start()
self.ready = True
self._start_time = time.time()
while self.ready:
try:
self.tick()
except (KeyboardInterrupt, SystemExit):
raise
except:
self.error_log("Error in HTTPServer.tick", level=logging.ERROR,
traceback=True)
if self.interrupt:
while self.interrupt is True:
# Wait for self.stop() to complete. See _set_interrupt.
time.sleep(0.1)
if self.interrupt:
raise self.interrupt
def error_log(self, msg="", level=20, traceback=False):
# Override this in subclasses as desired
sys.stderr.write(msg + '\n')
sys.stderr.flush()
if traceback:
tblines = format_exc()
sys.stderr.write(tblines)
sys.stderr.flush()
def bind(self, family, type, proto=0):
"""Create (or recreate) the actual socket object."""
self.socket = socket.socket(family, type, proto)
prevent_socket_inheritance(self.socket)
self.socket.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
if self.nodelay and not isinstance(self.bind_addr, str):
self.socket.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)
if self.ssl_adapter is not None:
self.socket = self.ssl_adapter.bind(self.socket)
# If listening on the IPV6 any address ('::' = IN6ADDR_ANY),
# activate dual-stack. See
# https://bitbucket.org/cherrypy/cherrypy/issue/871.
if (hasattr(socket, 'AF_INET6') and family == socket.AF_INET6
and self.bind_addr[0] in ('::', '::0', '::0.0.0.0')):
try:
self.socket.setsockopt(
socket.IPPROTO_IPV6, socket.IPV6_V6ONLY, 0)
except (AttributeError, socket.error):
# Apparently, the socket option is not available in
# this machine's TCP stack
pass
self.socket.bind(self.bind_addr)
def tick(self):
"""Accept a new connection and put it on the Queue."""
try:
s, addr = self.socket.accept()
if self.stats['Enabled']:
self.stats['Accepts'] += 1
if not self.ready:
return
prevent_socket_inheritance(s)
if hasattr(s, 'settimeout'):
s.settimeout(self.timeout)
makefile = CP_fileobject
ssl_env = {}
# if ssl cert and key are set, we try to be a secure HTTP server
if self.ssl_adapter is not None:
try:
s, ssl_env = self.ssl_adapter.wrap(s)
except NoSSLError:
msg = ("The client sent a plain HTTP request, but "
"this server only speaks HTTPS on this port.")
buf = ["%s 400 Bad Request\r\n" % self.protocol,
"Content-Length: %s\r\n" % len(msg),
"Content-Type: text/plain\r\n\r\n",
msg]
wfile = makefile(s._sock, "wb", DEFAULT_BUFFER_SIZE)
try:
wfile.sendall("".join(buf))
except socket.error:
x = sys.exc_info()[1]
if x.args[0] not in socket_errors_to_ignore:
raise
return
if not s:
return
makefile = self.ssl_adapter.makefile
# Re-apply our timeout since we may have a new socket object
if hasattr(s, 'settimeout'):
s.settimeout(self.timeout)
conn = self.ConnectionClass(self, s, makefile)
if not isinstance(self.bind_addr, basestring):
# optional values
# Until we do DNS lookups, omit REMOTE_HOST
if addr is None: # sometimes this can happen
# figure out if AF_INET or AF_INET6.
if len(s.getsockname()) == 2:
# AF_INET
addr = ('0.0.0.0', 0)
else:
# AF_INET6
addr = ('::', 0)
conn.remote_addr = addr[0]
conn.remote_port = addr[1]
conn.ssl_env = ssl_env
try:
self.requests.put(conn)
except queue.Full:
# Just drop the conn. TODO: write 503 back?
conn.close()
return
except socket.timeout:
# The only reason for the timeout in start() is so we can
# notice keyboard interrupts on Win32, which don't interrupt
# accept() by default
return
except socket.error:
x = sys.exc_info()[1]
if self.stats['Enabled']:
self.stats['Socket Errors'] += 1
if x.args[0] in socket_error_eintr:
# I *think* this is right. EINTR should occur when a signal
# is received during the accept() call; all docs say retry
# the call, and I *think* I'm reading it right that Python
# will then go ahead and poll for and handle the signal
# elsewhere. See
# https://bitbucket.org/cherrypy/cherrypy/issue/707.
return
if x.args[0] in socket_errors_nonblocking:
# Just try again. See
# https://bitbucket.org/cherrypy/cherrypy/issue/479.
return
if x.args[0] in socket_errors_to_ignore:
# Our socket was closed.
# See https://bitbucket.org/cherrypy/cherrypy/issue/686.
return
raise
def _get_interrupt(self):
return self._interrupt
def _set_interrupt(self, interrupt):
self._interrupt = True
self.stop()
self._interrupt = interrupt
interrupt = property(_get_interrupt, _set_interrupt,
doc="Set this to an Exception instance to "
"interrupt the server.")
def stop(self):
"""Gracefully shutdown a server that is serving forever."""
self.ready = False
if self._start_time is not None:
self._run_time += (time.time() - self._start_time)
self._start_time = None
sock = getattr(self, "socket", None)
if sock:
if not isinstance(self.bind_addr, basestring):
# Touch our own socket to make accept() return immediately.
try:
host, port = sock.getsockname()[:2]
except socket.error:
x = sys.exc_info()[1]
if x.args[0] not in socket_errors_to_ignore:
# Changed to use error code and not message
# See
# https://bitbucket.org/cherrypy/cherrypy/issue/860.
raise
else:
# Note that we're explicitly NOT using AI_PASSIVE,
# here, because we want an actual IP to touch.
# localhost won't work if we've bound to a public IP,
# but it will if we bound to '0.0.0.0' (INADDR_ANY).
for res in socket.getaddrinfo(host, port, socket.AF_UNSPEC,
socket.SOCK_STREAM):
af, socktype, proto, canonname, sa = res
s = None
try:
s = socket.socket(af, socktype, proto)
# See
# http://groups.google.com/group/cherrypy-users/
# browse_frm/thread/bbfe5eb39c904fe0
s.settimeout(1.0)
s.connect((host, port))
s.close()
except socket.error:
if s:
s.close()
if hasattr(sock, "close"):
sock.close()
self.socket = None
self.requests.stop(self.shutdown_timeout)
class Gateway(object):
"""A base class to interface HTTPServer with other systems, such as WSGI.
"""
def __init__(self, req):
self.req = req
def respond(self):
"""Process the current request. Must be overridden in a subclass."""
raise NotImplemented
# These may either be wsgiserver.SSLAdapter subclasses or the string names
# of such classes (in which case they will be lazily loaded).
ssl_adapters = {
'builtin': 'cherrypy.wsgiserver.ssl_builtin.BuiltinSSLAdapter',
'pyopenssl': 'cherrypy.wsgiserver.ssl_pyopenssl.pyOpenSSLAdapter',
}
def get_ssl_adapter_class(name='pyopenssl'):
"""Return an SSL adapter class for the given name."""
adapter = ssl_adapters[name.lower()]
if isinstance(adapter, basestring):
last_dot = adapter.rfind(".")
attr_name = adapter[last_dot + 1:]
mod_path = adapter[:last_dot]
try:
mod = sys.modules[mod_path]
if mod is None:
raise KeyError()
except KeyError:
# The last [''] is important.
mod = __import__(mod_path, globals(), locals(), [''])
# Let an AttributeError propagate outward.
try:
adapter = getattr(mod, attr_name)
except AttributeError:
raise AttributeError("'%s' object has no attribute '%s'"
% (mod_path, attr_name))
return adapter
# ------------------------------- WSGI Stuff -------------------------------- #
class CherryPyWSGIServer(HTTPServer):
"""A subclass of HTTPServer which calls a WSGI application."""
wsgi_version = (1, 0)
"""The version of WSGI to produce."""
def __init__(self, bind_addr, wsgi_app, numthreads=10, server_name=None,
max=-1, request_queue_size=5, timeout=10, shutdown_timeout=5,
accepted_queue_size=-1, accepted_queue_timeout=10):
self.requests = ThreadPool(self, min=numthreads or 1, max=max,
accepted_queue_size=accepted_queue_size,
accepted_queue_timeout=accepted_queue_timeout)
self.wsgi_app = wsgi_app
self.gateway = wsgi_gateways[self.wsgi_version]
self.bind_addr = bind_addr
if not server_name:
server_name = socket.gethostname()
self.server_name = server_name
self.request_queue_size = request_queue_size
self.timeout = timeout
self.shutdown_timeout = shutdown_timeout
self.clear_stats()
def _get_numthreads(self):
return self.requests.min
def _set_numthreads(self, value):
self.requests.min = value
numthreads = property(_get_numthreads, _set_numthreads)
class WSGIGateway(Gateway):
"""A base class to interface HTTPServer with WSGI."""
def __init__(self, req):
self.req = req
self.started_response = False
self.env = self.get_environ()
self.remaining_bytes_out = None
def get_environ(self):
"""Return a new environ dict targeting the given wsgi.version"""
raise NotImplemented
def respond(self):
"""Process the current request."""
response = self.req.server.wsgi_app(self.env, self.start_response)
try:
for chunk in response:
# "The start_response callable must not actually transmit
# the response headers. Instead, it must store them for the
# server or gateway to transmit only after the first
# iteration of the application return value that yields
# a NON-EMPTY string, or upon the application's first
# invocation of the write() callable." (PEP 333)
if chunk:
if isinstance(chunk, unicodestr):
chunk = chunk.encode('ISO-8859-1')
self.write(chunk)
finally:
if hasattr(response, "close"):
response.close()
def start_response(self, status, headers, exc_info=None):
"""WSGI callable to begin the HTTP response."""
# "The application may call start_response more than once,
# if and only if the exc_info argument is provided."
if self.started_response and not exc_info:
raise AssertionError("WSGI start_response called a second "
"time with no exc_info.")
self.started_response = True
# "if exc_info is provided, and the HTTP headers have already been
# sent, start_response must raise an error, and should raise the
# exc_info tuple."
if self.req.sent_headers:
try:
raise exc_info[0], exc_info[1], exc_info[2]
finally:
exc_info = None
self.req.status = status
for k, v in headers:
if not isinstance(k, str):
raise TypeError(
"WSGI response header key %r is not of type str." % k)
if not isinstance(v, str):
raise TypeError(
"WSGI response header value %r is not of type str." % v)
if k.lower() == 'content-length':
self.remaining_bytes_out = int(v)
self.req.outheaders.extend(headers)
return self.write
def write(self, chunk):
"""WSGI callable to write unbuffered data to the client.
This method is also used internally by start_response (to write
data from the iterable returned by the WSGI application).
"""
if not self.started_response:
raise AssertionError("WSGI write called before start_response.")
chunklen = len(chunk)
rbo = self.remaining_bytes_out
if rbo is not None and chunklen > rbo:
if not self.req.sent_headers:
# Whew. We can send a 500 to the client.
self.req.simple_response(
"500 Internal Server Error",
"The requested resource returned more bytes than the "
"declared Content-Length.")
else:
# Dang. We have probably already sent data. Truncate the chunk
# to fit (so the client doesn't hang) and raise an error later.
chunk = chunk[:rbo]
if not self.req.sent_headers:
self.req.sent_headers = True
self.req.send_headers()
self.req.write(chunk)
if rbo is not None:
rbo -= chunklen
if rbo < 0:
raise ValueError(
"Response body exceeds the declared Content-Length.")
class WSGIGateway_10(WSGIGateway):
"""A Gateway class to interface HTTPServer with WSGI 1.0.x."""
def get_environ(self):
"""Return a new environ dict targeting the given wsgi.version"""
req = self.req
env = {
# set a non-standard environ entry so the WSGI app can know what
# the *real* server protocol is (and what features to support).
# See http://www.faqs.org/rfcs/rfc2145.html.
'ACTUAL_SERVER_PROTOCOL': req.server.protocol,
'PATH_INFO': req.path,
'QUERY_STRING': req.qs,
'REMOTE_ADDR': req.conn.remote_addr or '',
'REMOTE_PORT': str(req.conn.remote_port or ''),
'REQUEST_METHOD': req.method,
'REQUEST_URI': req.uri,
'SCRIPT_NAME': '',
'SERVER_NAME': req.server.server_name,
# Bah. "SERVER_PROTOCOL" is actually the REQUEST protocol.
'SERVER_PROTOCOL': req.request_protocol,
'SERVER_SOFTWARE': req.server.software,
'wsgi.errors': sys.stderr,
'wsgi.input': req.rfile,
'wsgi.multiprocess': False,
'wsgi.multithread': True,
'wsgi.run_once': False,
'wsgi.url_scheme': req.scheme,
'wsgi.version': (1, 0),
}
if isinstance(req.server.bind_addr, basestring):
# AF_UNIX. This isn't really allowed by WSGI, which doesn't
# address unix domain sockets. But it's better than nothing.
env["SERVER_PORT"] = ""
else:
env["SERVER_PORT"] = str(req.server.bind_addr[1])
# Request headers
for k, v in req.inheaders.iteritems():
env["HTTP_" + k.upper().replace("-", "_")] = v
# CONTENT_TYPE/CONTENT_LENGTH
ct = env.pop("HTTP_CONTENT_TYPE", None)
if ct is not None:
env["CONTENT_TYPE"] = ct
cl = env.pop("HTTP_CONTENT_LENGTH", None)
if cl is not None:
env["CONTENT_LENGTH"] = cl
if req.conn.ssl_env:
env.update(req.conn.ssl_env)
return env
class WSGIGateway_u0(WSGIGateway_10):
"""A Gateway class to interface HTTPServer with WSGI u.0.
WSGI u.0 is an experimental protocol, which uses unicode for keys and
values in both Python 2 and Python 3.
"""
def get_environ(self):
"""Return a new environ dict targeting the given wsgi.version"""
req = self.req
env_10 = WSGIGateway_10.get_environ(self)
env = dict([(k.decode('ISO-8859-1'), v)
for k, v in env_10.iteritems()])
env[u'wsgi.version'] = ('u', 0)
# Request-URI
env.setdefault(u'wsgi.url_encoding', u'utf-8')
try:
for key in [u"PATH_INFO", u"SCRIPT_NAME", u"QUERY_STRING"]:
env[key] = env_10[str(key)].decode(env[u'wsgi.url_encoding'])
except UnicodeDecodeError:
# Fall back to latin 1 so apps can transcode if needed.
env[u'wsgi.url_encoding'] = u'ISO-8859-1'
for key in [u"PATH_INFO", u"SCRIPT_NAME", u"QUERY_STRING"]:
env[key] = env_10[str(key)].decode(env[u'wsgi.url_encoding'])
for k, v in sorted(env.items()):
if isinstance(v, str) and k not in ('REQUEST_URI', 'wsgi.input'):
env[k] = v.decode('ISO-8859-1')
return env
wsgi_gateways = {
(1, 0): WSGIGateway_10,
('u', 0): WSGIGateway_u0,
}
class WSGIPathInfoDispatcher(object):
"""A WSGI dispatcher for dispatch based on the PATH_INFO.
apps: a dict or list of (path_prefix, app) pairs.
"""
def __init__(self, apps):
try:
apps = list(apps.items())
except AttributeError:
pass
# Sort the apps by len(path), descending
apps.sort(cmp=lambda x, y: cmp(len(x[0]), len(y[0])))
apps.reverse()
# The path_prefix strings must start, but not end, with a slash.
# Use "" instead of "/".
self.apps = [(p.rstrip("/"), a) for p, a in apps]
def __call__(self, environ, start_response):
path = environ["PATH_INFO"] or "/"
for p, app in self.apps:
# The apps list should be sorted by length, descending.
if path.startswith(p + "/") or path == p:
environ = environ.copy()
environ["SCRIPT_NAME"] = environ["SCRIPT_NAME"] + p
environ["PATH_INFO"] = path[len(p):]
return app(environ, start_response)
start_response('404 Not Found', [('Content-Type', 'text/plain'),
('Content-Length', '0')])
return ['']
|
friture | scope_data | #!/usr/bin/env python
# -*- coding: utf-8 -*-
# Copyright (C) 2021 Timothée Lecomte
# This file is part of Friture.
#
# Friture is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License version 3 as published by
# the Free Software Foundation.
#
# Friture is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Friture. If not, see <http://www.gnu.org/licenses/>.
from friture.axis import Axis
from friture.curve import Curve
from PyQt5 import QtCore
from PyQt5.QtCore import pyqtProperty
from PyQt5.QtQml import QQmlListProperty
class Scope_Data(QtCore.QObject):
show_legend_changed = QtCore.pyqtSignal(bool)
plot_items_changed = QtCore.pyqtSignal()
def __init__(self, parent=None):
super().__init__(parent)
self._plot_items = []
self._horizontal_axis = Axis(self)
self._vertical_axis = Axis(self)
self._show_legend = True
@pyqtProperty(QQmlListProperty, notify=plot_items_changed)
def plot_items(self):
return QQmlListProperty(Curve, self, self._plot_items)
def insert_plot_item(self, index, plot_item):
self._plot_items.insert(index, plot_item)
self.plot_items_changed.emit()
def add_plot_item(self, plot_item):
self._plot_items.append(plot_item)
plot_item.setParent(self) # take ownership
self.plot_items_changed.emit()
def remove_plot_item(self, plot_item):
self._plot_items.remove(plot_item)
self.plot_items_changed.emit()
@pyqtProperty(Axis, constant=True)
def horizontal_axis(self):
return self._horizontal_axis
@pyqtProperty(Axis, constant=True)
def vertical_axis(self):
return self._vertical_axis
@pyqtProperty(bool, notify=show_legend_changed)
def show_legend(self):
return self._show_legend
@show_legend.setter
def show_legend(self, show_legend):
if self._show_legend != show_legend:
self._show_legend = show_legend
self.show_legend_changed.emit(show_legend)
|
ipv8 | ipv8_component | from typing import Optional
from ipv8.bootstrapping.dispersy.bootstrapper import DispersyBootstrapper
from ipv8.configuration import DISPERSY_BOOTSTRAPPER, ConfigBuilder
from ipv8.dht.churn import PingChurn
from ipv8.dht.discovery import DHTDiscoveryCommunity
from ipv8.dht.routing import RoutingTable
from ipv8.messaging.interfaces.dispatcher.endpoint import DispatcherEndpoint
from ipv8.messaging.interfaces.udp.endpoint import UDPv4Address
from ipv8.peer import Peer
from ipv8.peerdiscovery.churn import RandomChurn
from ipv8.peerdiscovery.community import DiscoveryCommunity, PeriodicSimilarity
from ipv8.peerdiscovery.discovery import RandomWalk
from ipv8.taskmanager import TaskManager
from ipv8_service import IPv8
from tribler.core.components.component import Component
from tribler.core.components.key.key_component import KeyComponent
INFINITE = -1
# pylint: disable=import-outside-toplevel
class Ipv8Component(Component):
ipv8: IPv8 = None
peer: Peer
dht_discovery_community: Optional[DHTDiscoveryCommunity] = None
_task_manager: TaskManager
_peer_discovery_community: Optional[DiscoveryCommunity] = None
async def run(self):
await super().run()
config = self.session.config
self._task_manager = TaskManager()
port = config.ipv8.port
address = config.ipv8.address
self.logger.info("Starting ipv8")
self.logger.info(f"Port: {port}. Address: {address}")
ipv8_config_builder = (
ConfigBuilder()
.set_port(port)
.set_address(address)
.clear_overlays()
.clear_keys() # We load the keys ourselves
.set_working_directory(str(config.state_dir))
.set_walker_interval(config.ipv8.walk_interval)
)
if config.gui_test_mode:
endpoint = DispatcherEndpoint([])
else:
# IPv8 includes IPv6 support by default.
# We only load IPv4 to not kill all Tribler overlays (currently, it would instantly crash all users).
# If you want to test IPv6 in Tribler you can set ``endpoint = None`` here.
endpoint = DispatcherEndpoint(
["UDPIPv4"], UDPIPv4={"port": port, "ip": address}
)
ipv8 = IPv8(
ipv8_config_builder.finalize(),
enable_statistics=config.ipv8.statistics and not config.gui_test_mode,
endpoint_override=endpoint,
)
await ipv8.start()
self.ipv8 = ipv8
key_component = await self.require_component(KeyComponent)
self.peer = Peer(key_component.primary_key)
if config.ipv8.walk_scaling_enabled and not config.gui_test_mode:
from tribler.core.components.ipv8.ipv8_health_monitor import IPv8Monitor
IPv8Monitor(
ipv8, config.ipv8.walk_interval, config.ipv8.walk_scaling_upper_limit
).start(self._task_manager)
if config.dht.enabled:
self._init_dht_discovery_community()
if not config.gui_test_mode:
if config.discovery_community.enabled:
self._init_peer_discovery_community()
else:
if config.dht.enabled:
self.dht_discovery_community.routing_tables[
UDPv4Address
] = RoutingTable("\x00" * 20)
def initialise_community_by_default(
self, community, default_random_walk_max_peers=20
):
community.bootstrappers.append(self.make_bootstrapper())
# Value of `target_peers` must not be equal to the value of `max_peers` for the community.
# This causes a deformed network topology and makes it harder for peers to connect to others.
# More information: https://github.com/Tribler/py-ipv8/issues/979#issuecomment-896643760
#
# Then:
# random_walk_max_peers should be less than community.max_peers:
random_walk_max_peers = min(
default_random_walk_max_peers, community.max_peers - 10
)
# random_walk_max_peers should be greater than 0
random_walk_max_peers = max(0, random_walk_max_peers)
self.ipv8.add_strategy(community, RandomWalk(community), random_walk_max_peers)
if (
self.session.config.ipv8.statistics
and not self.session.config.gui_test_mode
):
self.ipv8.endpoint.enable_community_statistics(community.get_prefix(), True)
async def unload_community(self, community):
await self.ipv8.unload_overlay(community)
def make_bootstrapper(self) -> DispersyBootstrapper:
args = DISPERSY_BOOTSTRAPPER["init"]
if bootstrap_override := self.session.config.ipv8.bootstrap_override:
address, port = bootstrap_override.split(":")
args = {"ip_addresses": [(address, int(port))], "dns_addresses": []}
return DispersyBootstrapper(**args)
def _init_peer_discovery_community(self):
ipv8 = self.ipv8
community = DiscoveryCommunity(
self.peer, ipv8.endpoint, ipv8.network, max_peers=100
)
self.initialise_community_by_default(community)
ipv8.add_strategy(community, RandomChurn(community), INFINITE)
ipv8.add_strategy(community, PeriodicSimilarity(community), INFINITE)
self._peer_discovery_community = community
def _init_dht_discovery_community(self):
ipv8 = self.ipv8
community = DHTDiscoveryCommunity(
self.peer, ipv8.endpoint, ipv8.network, max_peers=60
)
self.initialise_community_by_default(community)
ipv8.add_strategy(community, PingChurn(community), INFINITE)
self.dht_discovery_community = community
async def shutdown(self):
await super().shutdown()
if not self.ipv8:
return
for overlay in (self.dht_discovery_community, self._peer_discovery_community):
if overlay:
await self.ipv8.unload_overlay(overlay)
await self._task_manager.shutdown_task_manager()
await self.ipv8.stop(stop_loop=False)
|
SysML | diagramtype | from gaphor.diagram.diagramtoolbox import DiagramType
from gaphor.diagram.group import change_owner
from gaphor.SysML.sysml import SysMLDiagram
class DiagramDefault:
def __init__(self, from_type, to_type, name):
self.from_type = from_type
self.to_type = to_type
self.name = name
class SysMLDiagramType(DiagramType):
def __init__(self, id, name, sections, allowed_types=(), defaults=()):
super().__init__(id, name, sections)
self._allowed_types = allowed_types
assert all(d.to_type in allowed_types for d in defaults)
self._defaults = defaults
def allowed(self, element) -> bool:
if isinstance(element, self._allowed_types):
return True
return any(isinstance(element, d.from_type) for d in self._defaults)
def create(self, element_factory, element):
if not isinstance(element, self._allowed_types):
d = next(d for d in self._defaults if isinstance(element, d.from_type))
assert d.to_type in self._allowed_types
new_element = element_factory.create(d.to_type)
new_element.name = d.name
if element:
change_owner(element, new_element)
element = new_element
diagram = element_factory.create(SysMLDiagram)
diagram.name = diagram.gettext(self.name)
diagram.diagramType = self.id
if element:
change_owner(element, diagram)
return diagram
|
library | file | # This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
import os
import time
from pathlib import Path
from typing import Dict, Generator, Iterable, Optional, Set, Tuple, Union
from gi.repository import Gio, GLib, GObject
from quodlibet import _, formats, print_d, print_w
from quodlibet.formats import AudioFile, AudioFileError
from quodlibet.library.base import Library, PicklingMixin, iter_paths
from quodlibet.qltk.notif import Task
from quodlibet.util import copool, print_exc
from quodlibet.util.library import get_exclude_dirs
from quodlibet.util.path import ismount, normalize_path, unexpand
from senf import fsn2text, fsnative
class FileLibrary(Library[fsnative, AudioFile], PicklingMixin):
"""A library containing items on a local(-ish) filesystem.
These must support the valid, exists, mounted, and reload methods,
and have a mountpoint attribute.
"""
def __init__(self, name=None):
super().__init__(name)
self._masked = {}
def _load_init(self, items):
"""Add many items to the library, check if the
mountpoints are available and mark items as masked if not.
Does not check if items are valid.
"""
mounts = {}
contents = self._contents
masked = self._masked
for item in items:
mountpoint = item.mountpoint
if mountpoint not in mounts:
is_mounted = ismount(mountpoint)
# In case mountpoint is mounted through autofs we need to
# access a sub path for it to mount
# https://github.com/quodlibet/quodlibet/issues/2146
if not is_mounted:
item.exists()
is_mounted = ismount(mountpoint)
mounts[mountpoint] = is_mounted
# at least one not mounted, make sure masked has an entry
if not is_mounted:
masked.setdefault(mountpoint, {})
if mounts[mountpoint]:
contents[item.key] = item
else:
masked[mountpoint][item.key] = item
def _load_item(self, item, force=False):
"""Add an item, or refresh it if it's already in the library.
No signals will be fired.
Return a tuple of booleans: (changed, removed)
"""
print_d(f"Loading {item.key!r}", self._name)
valid = item.valid()
# The item is fine; add it if it's not present.
if not force and valid:
print_d(f"{item.key!r} is valid.", self._name)
self._contents[item.key] = item
return False, False
else:
# Either we should force a load, or the item is not okay.
# We're going to reload; this could change the key. So
# remove the item if it's currently in.
try:
del self._contents[item.key]
except KeyError:
present = False
else:
present = True
# If the item still exists, reload it.
if item.exists():
try:
item.reload()
except AudioFileError:
print_w(f"Error reloading {item.key!r}", self._name)
return False, True
else:
print_d(f"Reloaded {item.key!r}.", self._name)
self._contents[item.key] = item
return True, False
elif not item.mounted():
# We don't know if the item is okay or not, since
# it's not not mounted. If the item was present
# we need to mark it as removed.
print_d(f"Masking {item.key!r}", self._name)
self._masked.setdefault(item.mountpoint, {})
self._masked[item.mountpoint][item.key] = item
return False, present
else:
# The item doesn't exist at all anymore. Mark it as
# removed if it was present, otherwise nothing.
print_d(f"Ignoring (so removing) {item.key!r}.", self._name)
return False, present
def reload(self, item, changed=None, removed=None):
"""Reload a song, possibly noting its status.
If sets are given, it assumes the caller will handle signals,
and only updates the sets. Otherwise, it handles signals
itself. It *always* handles library contents, so do not
try to remove (again) a song that appears in the removed set.
"""
was_changed, was_removed = self._load_item(item, force=True)
assert not (was_changed and was_removed)
if was_changed:
if changed is None:
self.emit("changed", {item})
else:
changed.add(item)
elif was_removed:
if removed is None:
self.emit("removed", {item})
else:
removed.add(item)
def rebuild(self, paths, force=False, exclude=None, cofuncid=None):
"""Reload or remove songs if they have changed or been deleted.
This generator rebuilds the library over the course of iteration.
Any paths given will be scanned for new files, using the 'scan'
method.
Only items present in the library when the rebuild is started
will be checked.
If this function is copooled, set "cofuncid" to enable pause/stop
buttons in the UI.
"""
print_d(f"Rebuilding, force is {force}", self._name)
task = Task(_("Library"), _("Checking mount points"))
if cofuncid:
task.copool(cofuncid)
for i, (point, items) in task.list(enumerate(self._masked.items())):
if ismount(point):
self._contents.update(items)
del self._masked[point]
self.emit("added", list(items.values()))
yield True
task = Task(_("Library"), _("Scanning library"))
if cofuncid:
task.copool(cofuncid)
changed, removed = set(), set()
for i, (key, item) in task.list(enumerate(sorted(self.items()))):
if key in self._contents and force or not item.valid():
self.reload(item, changed, removed)
# These numbers are pretty empirical. We should yield more
# often than we emit signals; that way the main loop stays
# interactive and doesn't get bogged down in updates.
if len(changed) >= 200:
self.emit("changed", changed)
changed = set()
if len(removed) >= 200:
self.emit("removed", removed)
removed = set()
if len(changed) > 20 or i % 200 == 0:
yield True
print_d(f"Removing {len(removed)}, changing {len(changed)}).", self._name)
if removed:
self.emit("removed", removed)
if changed:
self.emit("changed", changed)
for value in self.scan(paths, exclude, cofuncid):
yield value
def add_filename(
self, filename: Union[str, Path], add: bool = True
) -> Optional[AudioFile]:
"""Add a file based on its filename.
Subclasses must override this to open the file correctly.
:return: the audio file if added (or None)
"""
pass
def contains_filename(self, filename) -> bool:
"""Returns if a song for the passed filename is in the library."""
key = normalize_path(filename, True)
return key in self._contents
def scan(
self,
paths: Iterable[fsnative],
exclude: Optional[Iterable[fsnative]] = None,
cofuncid=None,
):
def need_yield(last_yield=[0]):
current = time.time()
if abs(current - last_yield[0]) > 0.015:
last_yield[0] = current
return True
return False
def need_added(last_added=[0]):
current = time.time()
if abs(current - last_added[0]) > 1.0:
last_added[0] = current
return True
return False
# first scan each path for new files
paths_to_load = []
for scan_path in paths:
print_d(f"Scanning {scan_path}", self._name)
desc = _("Scanning %s") % (fsn2text(unexpand(scan_path)))
with Task(_("Library"), desc) as task:
if cofuncid:
task.copool(cofuncid)
for real_path in iter_paths(scan_path, exclude=exclude):
if need_yield():
task.pulse()
yield
# skip unknown file extensions
if not formats.filter(real_path):
continue
# already loaded
if self.contains_filename(real_path):
continue
paths_to_load.append(real_path)
yield
# then (try to) load all new files
with Task(_("Library"), _("Loading files")) as task:
if cofuncid:
task.copool(cofuncid)
added = []
for real_path in task.gen(paths_to_load):
item = self.add_filename(real_path, False)
if item is not None:
added.append(item)
if len(added) > 100 or need_added():
self.add(added)
added = []
yield
if added and need_yield():
yield
if added:
self.add(added)
added = []
yield True
def get_content(self):
"""Return visible and masked items"""
items = list(self.values())
for masked in self._masked.values():
items.extend(masked.values())
# Item keys are often based on filenames, in which case
# sorting takes advantage of the filesystem cache when we
# reload/rescan the files.
items.sort(key=lambda item: item.key)
return items
def masked(self, item):
"""Return true if the item is in the library but masked."""
try:
point = item.mountpoint
except AttributeError:
# Checking a key.
for point in self._masked.values():
if item in point:
return True
else:
# Checking a full item.
return item in self._masked.get(point, {}).values()
def unmask(self, point):
print_d(f"Unmasking {point!r}", self._name)
items = self._masked.pop(point, {})
if items:
self.add(items.values())
def mask(self, point):
print_d(f"Masking {point!r}", self._name)
removed = {}
for item in self.values():
if item.mountpoint == point:
removed[item.key] = item
if removed:
self.remove(removed.values())
self._masked.setdefault(point, {}).update(removed)
@property
def masked_mount_points(self):
"""List of mount points that contain masked items"""
return list(self._masked.keys())
def get_masked(self, mount_point):
"""List of items for a mount point"""
return list(self._masked.get(mount_point, {}).values())
def remove_masked(self, mount_point):
"""Remove all songs for a masked point"""
self._masked.pop(mount_point, {})
def move_root(
self, old_root: str, new_root: fsnative, write_files: bool = True
) -> Generator[None, None, None]:
"""
Move the root for all songs in a given (scan) directory.
We avoid dereferencing the destination, to allow users things like:
1. Symlink new_path -> old_root
2. Move QL root to new_path
3. Remove symlink
4. Move audio files: old_root -> new_path
"""
# TODO: only move primary library
old_path = Path(normalize_path(old_root, canonicalise=True)).expanduser()
new_path = Path(normalize_path(new_root)).expanduser()
if not old_path.is_dir():
print_w(
f"Source dir {str(old_path)!r} doesn't exist, assuming that's OK",
self._name,
)
if not new_path.is_dir():
raise ValueError(f"Destination {new_path!r} is not a directory")
print_d(f"Checking entire library for {str(old_path)!r}", self._name)
missing: Set[AudioFile] = set()
changed = set()
total = len(self)
if not total:
return
with Task(_("Library"), _("Moving library files")) as task:
yield
for i, song in enumerate(list(self.values())):
task.update(i / total)
key = normalize_path(song.key)
path = Path(key)
if old_path in path.parents:
# TODO: more Pathlib-friendly dir replacement...
new_key = key.replace(str(old_path), str(new_path), 1)
new_key = normalize_path(new_key, canonicalise=False)
if new_key == key:
print_w(f"Substitution failed for {key!r}", self._name)
# We need to update ~filename and ~mountpoint
song.sanitize()
if write_files:
song.write()
if self.move_song(song, new_key):
changed.add(song)
else:
missing.add(song)
elif not (i % 1000):
print_d(f"Not moved, for example: {key!r}", self._name)
if not i % 100:
yield
self.changed(changed)
if missing:
print_w(
f"Couldn't find {len(list(missing))} files: {missing}", self._name
)
yield
self.save()
print_d(
f"Done moving {len(changed)} track(s) (of {total}) "
f"to {str(new_path)!r}.",
self._name,
)
def move_song(self, song: AudioFile, new_path: fsnative) -> bool:
"""Updates the location of a song, without touching the file.
:returns: True if it was could be found (and moved)
"""
existed = True
key = song.key
print_d(f"Moving {key!r} -> {new_path!r}", self._name)
try:
del self._contents[key] # type: ignore
except KeyError:
existed = False
# Continue - maybe it's already moved
song.sanitize(new_path)
self._contents[new_path] = song
return existed
def remove_roots(self, old_roots: Iterable[str]) -> Generator[None, None, None]:
"""Remove library roots (scandirs) entirely, and all their songs"""
old_paths = [
Path(normalize_path(root, canonicalise=True)).expanduser()
for root in old_roots
]
total = len(self)
removed = set()
print_d(f"Removing library roots {old_roots}", self._name)
yield
with Task(_("Library"), _("Removing library files")) as task:
for i, song in enumerate(list(self.values())):
task.update(i / total)
key = normalize_path(song.key)
song_path = Path(key)
if any(path in song_path.parents for path in old_paths):
removed.add(song)
if not i % 100:
yield
if removed:
self.remove(removed)
else:
print_d(f"No tracks in {old_roots} to remove", self._name)
Event = Gio.FileMonitorEvent
class WatchedFileLibraryMixin(FileLibrary):
"""A File Library that sets up monitors on directories at refresh
and handles changes sensibly"""
def __init__(self, name=None):
super().__init__(name)
self._monitors: Dict[Path, Tuple[GObject.GObject, int]] = {}
print_d(f"Initialised {self!r}")
def monitor_dir(self, path: Path) -> None:
"""Monitors a single directory"""
# Only add one monitor per absolute path...
if path not in self._monitors:
f = Gio.File.new_for_path(str(path))
try:
monitor = f.monitor_directory(Gio.FileMonitorFlags.WATCH_MOVES, None)
except GLib.GError as e:
print_w(f"Couldn't watch {path} ({e})", self._name)
monitor = None
if not monitor:
return
handler_id = monitor.connect("changed", self.__file_changed)
# Don't destroy references - http://stackoverflow.com/q/4535227
self._monitors[path] = (monitor, handler_id)
print_d(f"Monitoring {path!s}", self._name)
def __file_changed(
self,
_monitor,
main_file: Gio.File,
other_file: Optional[Gio.File],
event: Gio.FileMonitorEvent,
) -> None:
if event == Event.CHANGES_DONE_HINT:
# This seems to work fine on most Linux, but not on Windows / macOS
# Or at least, not in CI anyway.
# So shortcut the whole thing
return
try:
file_path = main_file.get_path()
if file_path is None:
return
file_path = normalize_path(file_path, True)
song = self.get(file_path)
file_path = Path(file_path)
other_path = (
Path(normalize_path(other_file.get_path(), True))
if other_file
else None
)
if event in (Event.CREATED, Event.MOVED_IN):
if file_path.is_dir():
self.monitor_dir(file_path)
copool.add(self.scan, [str(file_path)])
elif not song:
print_d(f"Auto-adding created file: {file_path}", self._name)
self.add_filename(str(file_path))
elif event == Event.RENAMED:
if not other_path:
print_w(
f"No destination found for rename of {file_path}", self._name
)
if song:
print_d(f"Moving {file_path} to {other_path}...", self._name)
if self.move_song(song, str(other_path)): # type:ignore
print_w(f"Song {file_path} has gone")
elif self.is_monitored_dir(file_path):
if self.librarian:
print_d(
f"Moving tracks from {file_path} -> {other_path}...",
self._name,
)
copool.add(
self.librarian.move_root,
str(file_path),
str(other_path),
write_files=False,
priority=GLib.PRIORITY_DEFAULT,
)
self.unmonitor_dir(file_path)
if other_path:
self.monitor_dir(other_path)
else:
if other_path:
print_w(
f"Seems {file_path} is not a track (deleted?)", self._name
)
# On some (Windows?) systems CHANGED is called which can remove
# before we get here, so let's try adding the new path back
self.add_filename(other_path)
elif event == Event.CHANGED:
if song:
# QL created (or knew about) this one; still check if it changed
if not song.valid():
self.reload(song)
else:
print_d(f"Auto-adding new file: {file_path}", self._name)
self.add_filename(file_path)
elif event in (Event.MOVED_OUT, Event.DELETED):
if song:
print_d(f"...so deleting {file_path}", self._name)
self.reload(song)
else:
# either not a song, or a song that was renamed by QL
if self.is_monitored_dir(file_path):
self.unmonitor_dir(file_path)
# And try to remove all songs under that dir. Slowly.
gone = set()
for key, song in self.iteritems():
if file_path in Path(key).parents:
gone.add(song)
if gone:
print_d(
f"Removing {len(gone)} contained songs in {file_path}",
self._name,
)
actually_gone = self.remove(gone)
if gone != actually_gone:
print_w(
f"Couldn't remove all: {gone - actually_gone}",
self._name,
)
else:
print_d(
f"Unhandled event {event} on {file_path} ({other_path})", self._name
)
return
except Exception:
print_w("Failed to run file monitor callback", self._name)
print_exc()
print_d(f"Finished handling {event}", self._name)
def is_monitored_dir(self, path: Path) -> bool:
return path in self._monitors
def unmonitor_dir(self, path: Path) -> None:
"""Disconnect and remove any monitor for a directory, if found"""
monitor, handler_id = self._monitors.get(path, (None, None))
if not monitor:
print_d(f"Couldn't find path {path} in active monitors", self._name)
return
monitor.disconnect(handler_id)
del self._monitors[path]
def start_watching(self, paths: Iterable[fsnative]):
print_d(f"Setting up file watches on {paths}...", self._name)
exclude_dirs = [e for e in get_exclude_dirs() if e]
def watching_producer():
# TODO: integrate this better with scanning.
for fullpath in paths:
desc = _("Adding watches for %s") % (fsn2text(unexpand(fullpath)))
with Task(_("Library"), desc) as task:
normalised = Path(normalize_path(fullpath, True)).expanduser()
if any(
Path(exclude) in normalised.parents for exclude in exclude_dirs
):
continue
unpulsed = 0
self.monitor_dir(normalised)
for path, dirs, files in os.walk(normalised):
normalised = Path(normalize_path(path, True))
for d in dirs:
self.monitor_dir(normalised / d)
unpulsed += len(dirs)
if unpulsed > 50:
task.pulse()
unpulsed = 0
yield
copool.add(watching_producer, funcid="watch_library")
def stop_watching(self):
print_d(f"Removing watches on {len(self._monitors)} dirs", self._name)
for monitor, handler_id in self._monitors.values():
monitor.disconnect(handler_id)
self._monitors.clear()
def destroy(self):
self.stop_watching()
super().destroy()
|
extractor | bostonglobe | # coding: utf-8
from __future__ import unicode_literals
import re
from ..utils import extract_attributes
from .common import InfoExtractor
class BostonGlobeIE(InfoExtractor):
_VALID_URL = (
r"(?i)https?://(?:www\.)?bostonglobe\.com/.*/(?P<id>[^/]+)/\w+(?:\.html)?"
)
_TESTS = [
{
"url": "http://www.bostonglobe.com/metro/2017/02/11/tree-finally-succumbs-disease-leaving-hole-neighborhood/h1b4lviqzMTIn9sVy8F3gP/story.html",
"md5": "0a62181079c85c2d2b618c9a738aedaf",
"info_dict": {
"title": "A tree finally succumbs to disease, leaving a hole in a neighborhood",
"id": "5320421710001",
"ext": "mp4",
"description": "It arrived as a sapling when the Back Bay was in its infancy, a spindly American elm tamped down into a square of dirt cut into the brick sidewalk of 1880s Marlborough Street, no higher than the first bay window of the new brownstone behind it.",
"timestamp": 1486877593,
"upload_date": "20170212",
"uploader_id": "245991542",
},
},
{
# Embedded youtube video; we hand it off to the Generic extractor.
"url": "https://www.bostonglobe.com/lifestyle/names/2017/02/17/does-ben-affleck-play-matt-damon-favorite-version-batman/ruqkc9VxKBYmh5txn1XhSI/story.html",
"md5": "582b40327089d5c0c949b3c54b13c24b",
"info_dict": {
"title": "Who Is Matt Damon's Favorite Batman?",
"id": "ZW1QCnlA6Qc",
"ext": "mp4",
"upload_date": "20170217",
"description": "md5:3b3dccb9375867e0b4d527ed87d307cb",
"uploader": "The Late Late Show with James Corden",
"uploader_id": "TheLateLateShow",
},
"expected_warnings": ["404"],
},
]
def _real_extract(self, url):
page_id = self._match_id(url)
webpage = self._download_webpage(url, page_id)
page_title = self._og_search_title(webpage, default=None)
# <video data-brightcove-video-id="5320421710001" data-account="245991542" data-player="SJWAiyYWg" data-embed="default" class="video-js" controls itemscope itemtype="http://schema.org/VideoObject">
entries = []
for video in re.findall(r"(?i)(<video[^>]+>)", webpage):
attrs = extract_attributes(video)
video_id = attrs.get("data-brightcove-video-id")
account_id = attrs.get("data-account")
player_id = attrs.get("data-player")
embed = attrs.get("data-embed")
if video_id and account_id and player_id and embed:
entries.append(
"http://players.brightcove.net/%s/%s_%s/index.html?videoId=%s"
% (account_id, player_id, embed, video_id)
)
if len(entries) == 0:
return self.url_result(url, "Generic")
elif len(entries) == 1:
return self.url_result(entries[0], "BrightcoveNew")
else:
return self.playlist_from_matches(
entries, page_id, page_title, ie="BrightcoveNew"
)
|
extractor | myvideoge | # coding: utf-8
from __future__ import unicode_literals
import re
from ..utils import (
MONTH_NAMES,
clean_html,
get_element_by_class,
get_element_by_id,
int_or_none,
js_to_json,
qualities,
unified_strdate,
)
from .common import InfoExtractor
class MyVideoGeIE(InfoExtractor):
_VALID_URL = r"https?://(?:www\.)?myvideo\.ge/v/(?P<id>[0-9]+)"
_TEST = {
"url": "https://www.myvideo.ge/v/3941048",
"md5": "8c192a7d2b15454ba4f29dc9c9a52ea9",
"info_dict": {
"id": "3941048",
"ext": "mp4",
"title": "The best prikol",
"upload_date": "20200611",
"thumbnail": r"re:^https?://.*\.jpg$",
"uploader": "chixa33",
"description": "md5:5b067801318e33c2e6eea4ab90b1fdd3",
},
# working from local dev system
"skip": "site blocks CI servers",
}
_MONTH_NAMES_KA = [
"იანვარი",
"თებერვალი",
"მარტი",
"აპრილი",
"მაისი",
"ივნისი",
"ივლისი",
"აგვისტო",
"სექტემბერი",
"ოქტომბერი",
"ნოემბერი",
"დეკემბერი",
]
_quality = staticmethod(qualities(("SD", "HD")))
def _real_extract(self, url):
video_id = self._match_id(url)
webpage = self._download_webpage(url, video_id)
title = (
self._og_search_title(webpage, default=None)
or clean_html(get_element_by_class("my_video_title", webpage))
or self._html_search_regex(
r"<title\b[^>]*>([^<]+)</title\b", webpage, "title"
)
)
jwplayer_sources = self._parse_json(
self._search_regex(
r"""(?s)jwplayer\s*\(\s*['"]mvplayer['"]\s*\)\s*\.\s*setup\s*\(.*?\bsources\s*:\s*(\[.*?])\s*[,});]""",
webpage,
"jwplayer sources",
fatal=False,
)
or "",
video_id,
transform_source=js_to_json,
fatal=False,
)
formats = self._parse_jwplayer_formats(jwplayer_sources or [], video_id)
for f in formats or []:
f["preference"] = self._quality(f["format_id"])
self._sort_formats(formats)
description = (
self._og_search_description(webpage)
or get_element_by_id("long_desc_holder", webpage)
or self._html_search_meta("description", webpage)
)
uploader = self._search_regex(
r'<a[^>]+class="mv_user_name"[^>]*>([^<]+)<',
webpage,
"uploader",
fatal=False,
)
upload_date = get_element_by_class("mv_vid_upl_date", webpage)
# as ka locale may not be present roll a local date conversion
upload_date = (
unified_strdate(
# translate any ka month to an en one
re.sub(
"|".join(self._MONTH_NAMES_KA),
lambda m: MONTH_NAMES["en"][self._MONTH_NAMES_KA.index(m.group(0))],
upload_date,
re.I,
)
)
if upload_date
else None
)
return {
"id": video_id,
"title": title,
"description": description,
"uploader": uploader,
"formats": formats,
"thumbnail": self._og_search_thumbnail(webpage),
"upload_date": upload_date,
"view_count": int_or_none(get_element_by_class("mv_vid_views", webpage)),
"like_count": int_or_none(get_element_by_id("likes_count", webpage)),
"dislike_count": int_or_none(get_element_by_id("dislikes_count", webpage)),
}
|
migrations | 0328_add_starter_feature_flag_template | from django.db import migrations
def create_starter_template(apps, schema_editor):
DashboardTemplate = apps.get_model("posthog", "DashboardTemplate")
DashboardTemplate.objects.create(
template_name="Flagged Feature Usage",
dashboard_description="Overview of engagement with the flagged feature including daily active users and weekly active users.",
dashboard_filters={},
tiles=[
{
"name": "Daily active users (DAUs)",
"type": "INSIGHT",
"color": "blue",
"filters": {
"events": ["{ENGAGEMENT}"],
"display": "ActionsLineGraph",
"insight": "TRENDS",
"interval": "day",
"date_from": "-30d",
},
"layouts": {
"sm": {"h": 5, "w": 6, "x": 0, "y": 0, "minH": 5, "minW": 3},
"xs": {"h": 5, "w": 1, "x": 0, "y": 0, "minH": 5, "minW": 3},
},
"description": "Shows the number of unique users that use your feature every day.",
},
{
"name": "Weekly active users (WAUs)",
"type": "INSIGHT",
"color": "green",
"filters": {
"events": ["{ENGAGEMENT}"],
"display": "ActionsLineGraph",
"insight": "TRENDS",
"interval": "week",
"date_from": "-90d",
},
"layouts": {
"sm": {"h": 5, "w": 6, "x": 6, "y": 0, "minH": 5, "minW": 3},
"xs": {"h": 5, "w": 1, "x": 0, "y": 5, "minH": 5, "minW": 3},
},
"description": "Shows the number of unique users that use your feature every week.",
},
],
tags=[],
variables=[
{
"id": "ENGAGEMENT",
"name": "Engagement",
"type": "event",
"default": {"name": "$pageview", "id": "$pageview"},
"required": True,
"description": "The event you use to define a user using the new feature",
}
],
scope="feature_flag",
)
class Migration(migrations.Migration):
dependencies = [
("posthog", "0327_alter_earlyaccessfeature_stage"),
]
operations = [
migrations.RunPython(
create_starter_template, reverse_code=migrations.RunPython.noop
)
]
|
extractor | xvideos | from __future__ import unicode_literals
import re
from ..compat import compat_urllib_parse_unquote
from ..utils import (
ExtractorError,
clean_html,
determine_ext,
int_or_none,
parse_duration,
)
from .common import InfoExtractor
class XVideosIE(InfoExtractor):
_VALID_URL = r"""(?x)
https?://
(?:
(?:[^/]+\.)?xvideos2?\.com/video|
(?:www\.)?xvideos\.es/video|
flashservice\.xvideos\.com/embedframe/|
static-hw\.xvideos\.com/swf/xv-player\.swf\?.*?\bid_video=
)
(?P<id>[0-9]+)
"""
_TESTS = [
{
"url": "http://www.xvideos.com/video4588838/biker_takes_his_girl",
"md5": "14cea69fcb84db54293b1e971466c2e1",
"info_dict": {
"id": "4588838",
"ext": "mp4",
"title": "Biker Takes his Girl",
"duration": 108,
"age_limit": 18,
},
},
{
"url": "https://flashservice.xvideos.com/embedframe/4588838",
"only_matching": True,
},
{
"url": "http://static-hw.xvideos.com/swf/xv-player.swf?id_video=4588838",
"only_matching": True,
},
{
"url": "http://xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://xvideos.es/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://www.xvideos.es/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "http://xvideos.es/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "http://www.xvideos.es/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "http://fr.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://fr.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "http://it.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://it.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "http://de.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
{
"url": "https://de.xvideos.com/video4588838/biker_takes_his_girl",
"only_matching": True,
},
]
def _real_extract(self, url):
video_id = self._match_id(url)
webpage = self._download_webpage(
"https://www.xvideos.com/video%s/0" % video_id, video_id
)
mobj = re.search(r'<h1 class="inlineError">(.+?)</h1>', webpage)
if mobj:
raise ExtractorError(
"%s said: %s" % (self.IE_NAME, clean_html(mobj.group(1))), expected=True
)
title = self._html_search_regex(
(
r"<title>(?P<title>.+?)\s+-\s+XVID",
r'setVideoTitle\s*\(\s*(["\'])(?P<title>(?:(?!\1).)+)\1',
),
webpage,
"title",
default=None,
group="title",
) or self._og_search_title(webpage)
thumbnails = []
for preference, thumbnail in enumerate(("", "169")):
thumbnail_url = self._search_regex(
r'setThumbUrl%s\(\s*(["\'])(?P<thumbnail>(?:(?!\1).)+)\1' % thumbnail,
webpage,
"thumbnail",
default=None,
group="thumbnail",
)
if thumbnail_url:
thumbnails.append(
{
"url": thumbnail_url,
"preference": preference,
}
)
duration = int_or_none(
self._og_search_property("duration", webpage, default=None)
) or parse_duration(
self._search_regex(
r'<span[^>]+class=["\']duration["\'][^>]*>.*?(\d[^<]+)',
webpage,
"duration",
fatal=False,
)
)
formats = []
video_url = compat_urllib_parse_unquote(
self._search_regex(r"flv_url=(.+?)&", webpage, "video URL", default="")
)
if video_url:
formats.append(
{
"url": video_url,
"format_id": "flv",
}
)
for kind, _, format_url in re.findall(
r'setVideo([^(]+)\((["\'])(http.+?)\2\)', webpage
):
format_id = kind.lower()
if format_id == "hls":
formats.extend(
self._extract_m3u8_formats(
format_url,
video_id,
"mp4",
entry_protocol="m3u8_native",
m3u8_id="hls",
fatal=False,
)
)
elif format_id in ("urllow", "urlhigh"):
formats.append(
{
"url": format_url,
"format_id": "%s-%s"
% (determine_ext(format_url, "mp4"), format_id[3:]),
"quality": -2 if format_id.endswith("low") else None,
}
)
self._sort_formats(formats)
return {
"id": video_id,
"formats": formats,
"title": title,
"duration": duration,
"thumbnails": thumbnails,
"age_limit": 18,
}
|
music-player | compile | #!/usr/bin/env python
import os
import sys
os.chdir(os.path.dirname(__file__))
# Import compile_utils from core.
sys.path += ["core"]
from glob import glob
import compile_utils as c
from compile_utils import *
# Compile core module.
sys_exec(["./core/compile.py"])
sys_exec(["cp", "core/musicplayer.so", "musicplayer.so"])
sys_exec(["mkdir", "-p", "build"])
os.chdir("build/")
# Compile _gui.so.
print("* Building _gui.so")
guiFiles = glob("../_gui/*.cpp")
cc(guiFiles, ["-I../core"] + get_python_ccopts())
link("../_gui.so", [c.get_cc_outfilename(fn) for fn in guiFiles], get_python_linkopts())
if sys.platform == "darwin":
# Compile _guiCocoa.so.
print("* Building _guiCocoa.so")
guiCocoaFiles = glob("../mac/gui/*.m*") + glob("../mac/gui/*.cpp")
PyObjCBridgeFile = "../mac/gui/PyObjCBridge.m"
guiCocoaFiles.remove(PyObjCBridgeFile) # we will handle that differently
cc(
guiCocoaFiles,
[
"-I../core",
"-I../_gui",
"-fobjc-arc",
]
+ get_python_ccopts(),
)
guiCocoaFiles.append(PyObjCBridgeFile)
cc(
[PyObjCBridgeFile],
[
"-I../core",
"-I../_gui",
"-I../mac/pyobjc-core/Modules/objc",
"-I../mac/pyobjc-core/libffi-src/include",
]
+ get_python_ccopts(),
)
link(
"../_guiCocoa.so",
[c.get_cc_outfilename(fn) for fn in guiCocoaFiles],
[
"-framework",
"Cocoa",
"-framework",
"Foundation",
"-framework",
"CoreFoundation",
]
+ get_python_linkopts(),
)
|
library | album | # This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
from quodlibet import print_d
from quodlibet.formats._audio import AlbumKey
from quodlibet.library.base import Library
from quodlibet.util.collection import Album
class AlbumLibrary(Library[AlbumKey, Album]):
"""An AlbumLibrary listens to a SongLibrary and sorts its songs into
albums.
The library behaves like a dictionary: the keys are album_keys of
AudioFiles, the values are Album objects.
"""
def __init__(self, library):
self.librarian = None
print_d("Initializing Album Library to watch %r" % library._name)
super().__init__("AlbumLibrary for %s" % library._name)
self._library = library
self._asig = library.connect("added", self.__added)
self._rsig = library.connect("removed", self.__removed)
self._csig = library.connect("changed", self.__changed)
self.__added(library, library.values(), signal=False)
def load(self):
# deprecated
pass
def destroy(self):
for sig in [self._asig, self._rsig, self._csig]:
self._library.disconnect(sig)
def _get(self, item):
return self._contents.get(item)
def __add(self, items):
changed = set()
new = set()
for song in items:
key = song.album_key
if key in self._contents:
changed.add(self._contents[key])
else:
album = Album(song)
self._contents[key] = album
new.add(album)
self._contents[key].songs.add(song)
changed -= new
return changed, new
def __added(self, library, items, signal=True):
changed, new = self.__add(items)
for album in changed:
album.finalize()
if signal:
if new:
self.emit("added", new)
if changed:
self.emit("changed", changed)
def __removed(self, library, items):
changed = set()
removed = set()
for song in items:
key = song.album_key
album = self._contents[key]
album.songs.remove(song)
changed.add(album)
if not album.songs:
removed.add(album)
del self._contents[key]
changed -= removed
for album in changed:
album.finalize()
if removed:
self.emit("removed", removed)
if changed:
self.emit("changed", changed)
def __changed(self, library, items):
"""Album keys could change between already existing ones... so we
have to do it the hard way and search by id."""
print_d("Updating affected albums for %d items" % len(items))
changed = set()
removed = set()
to_add = []
for song in items:
# in case the key hasn't changed
key = song.album_key
if key in self._contents and song in self._contents[key].songs:
changed.add(self._contents[key])
else: # key changed... look for it in each album
to_add.append(song)
for key, album in self._contents.items():
if song in album.songs:
album.songs.remove(song)
if not album.songs:
removed.add(album)
else:
changed.add(album)
break
# get new albums and changed ones because keys could have changed
add_changed, new = self.__add(to_add)
changed |= add_changed
# check if albums that were empty at some point are still empty
for album in removed:
if not album.songs:
del self._contents[album.key]
changed.discard(album)
for album in changed:
album.finalize()
if removed:
self.emit("removed", removed)
if changed:
self.emit("changed", changed)
if new:
self.emit("added", new)
|
const | attributes | # -*- coding: utf-8 -*-
# Automatically generated - don't edit.
# Use `python setup.py update_constants` to update it.
MB_ATTRIBUTES = {
"DB:cover_art_archive.art_type/name:001": "Front",
"DB:cover_art_archive.art_type/name:002": "Back",
"DB:cover_art_archive.art_type/name:003": "Booklet",
"DB:cover_art_archive.art_type/name:004": "Medium",
"DB:cover_art_archive.art_type/name:005": "Obi",
"DB:cover_art_archive.art_type/name:006": "Spine",
"DB:cover_art_archive.art_type/name:007": "Track",
"DB:cover_art_archive.art_type/name:008": "Other",
"DB:cover_art_archive.art_type/name:009": "Tray",
"DB:cover_art_archive.art_type/name:010": "Sticker",
"DB:cover_art_archive.art_type/name:011": "Poster",
"DB:cover_art_archive.art_type/name:012": "Liner",
"DB:cover_art_archive.art_type/name:013": "Watermark",
"DB:cover_art_archive.art_type/name:014": "Raw/Unedited",
"DB:cover_art_archive.art_type/name:015": "Matrix/Runout",
"DB:cover_art_archive.art_type/name:048": "Top",
"DB:cover_art_archive.art_type/name:049": "Bottom",
"DB:medium_format/name:001": "CD",
"DB:medium_format/name:002": "DVD",
"DB:medium_format/name:003": "SACD",
"DB:medium_format/name:004": "DualDisc",
"DB:medium_format/name:005": "LaserDisc",
"DB:medium_format/name:006": "MiniDisc",
"DB:medium_format/name:007": "Vinyl",
"DB:medium_format/name:008": "Cassette",
"DB:medium_format/name:009": "Cartridge",
"DB:medium_format/name:010": "Reel-to-reel",
"DB:medium_format/name:011": "DAT",
"DB:medium_format/name:012": "Digital Media",
"DB:medium_format/name:013": "Other",
"DB:medium_format/name:014": "Wax Cylinder",
"DB:medium_format/name:015": "Piano Roll",
"DB:medium_format/name:016": "DCC",
"DB:medium_format/name:017": "HD-DVD",
"DB:medium_format/name:018": "DVD-Audio",
"DB:medium_format/name:019": "DVD-Video",
"DB:medium_format/name:020": "Blu-ray",
"DB:medium_format/name:021": "VHS",
"DB:medium_format/name:022": "VCD",
"DB:medium_format/name:023": "SVCD",
"DB:medium_format/name:024": "Betamax",
"DB:medium_format/name:025": "HDCD",
"DB:medium_format/name:026": "USB Flash Drive",
"DB:medium_format/name:027": "slotMusic",
"DB:medium_format/name:028": "UMD",
"DB:medium_format/name:029": '7" Vinyl',
"DB:medium_format/name:030": '10" Vinyl',
"DB:medium_format/name:031": '12" Vinyl',
"DB:medium_format/name:033": "CD-R",
"DB:medium_format/name:034": "8cm CD",
"DB:medium_format/name:035": "Blu-spec CD",
"DB:medium_format/name:036": "SHM-CD",
"DB:medium_format/name:037": "HQCD",
"DB:medium_format/name:038": "Hybrid SACD",
"DB:medium_format/name:039": "CD+G",
"DB:medium_format/name:040": "8cm CD+G",
"DB:medium_format/name:041": "CDV",
"DB:medium_format/name:042": "Enhanced CD",
"DB:medium_format/name:043": "Data CD",
"DB:medium_format/name:044": "DTS CD",
"DB:medium_format/name:045": "Playbutton",
"DB:medium_format/name:046": "Download Card",
"DB:medium_format/name:047": "DVDplus",
"DB:medium_format/name:048": "VinylDisc",
"DB:medium_format/name:049": '3.5" Floppy Disk',
"DB:medium_format/name:050": "Edison Diamond Disc",
"DB:medium_format/name:051": "Flexi-disc",
"DB:medium_format/name:052": '7" Flexi-disc',
"DB:medium_format/name:053": "Shellac",
"DB:medium_format/name:054": '10" Shellac',
"DB:medium_format/name:055": '12" Shellac',
"DB:medium_format/name:056": '7" Shellac',
"DB:medium_format/name:057": "SHM-SACD",
"DB:medium_format/name:058": "Pathé disc",
"DB:medium_format/name:059": "VHD",
"DB:medium_format/name:060": "CED",
"DB:medium_format/name:061": "Copy Control CD",
"DB:medium_format/name:062": "SD Card",
"DB:medium_format/name:063": "Hybrid SACD (CD layer)",
"DB:medium_format/name:064": "Hybrid SACD (SACD layer)",
"DB:medium_format/name:065": "DualDisc (DVD-Audio side)",
"DB:medium_format/name:066": "DualDisc (DVD-Video side)",
"DB:medium_format/name:067": "DualDisc (CD side)",
"DB:medium_format/name:068": "DVDplus (DVD-Audio side)",
"DB:medium_format/name:069": "DVDplus (DVD-Video side)",
"DB:medium_format/name:070": "DVDplus (CD side)",
"DB:medium_format/name:071": '8" LaserDisc',
"DB:medium_format/name:072": '12" LaserDisc',
"DB:medium_format/name:073": "Phonograph record",
"DB:medium_format/name:074": "PlayTape",
"DB:medium_format/name:075": "HiPac",
"DB:medium_format/name:076": "Floppy Disk",
"DB:medium_format/name:077": "Zip Disk",
"DB:medium_format/name:078": "8-Track Cartridge",
"DB:medium_format/name:079": "Blu-ray-R",
"DB:medium_format/name:080": "VinylDisc (DVD side)",
"DB:medium_format/name:081": "VinylDisc (Vinyl side)",
"DB:medium_format/name:082": "VinylDisc (CD side)",
"DB:medium_format/name:083": "Microcassette",
"DB:medium_format/name:084": "SACD (2 channels)",
"DB:medium_format/name:085": "SACD (multichannel)",
"DB:medium_format/name:086": "Hybrid SACD (SACD layer, multichannel)",
"DB:medium_format/name:087": "Hybrid SACD (SACD layer, 2 channels)",
"DB:medium_format/name:088": "SHM-SACD (multichannel)",
"DB:medium_format/name:089": "SHM-SACD (2 channels)",
"DB:medium_format/name:090": "Tefifon",
"DB:medium_format/name:091": '5.25" Floppy Disk',
"DB:medium_format/name:092": "DVD-R Video",
"DB:medium_format/name:093": "Data DVD-R",
"DB:medium_format/name:094": "Data DVD",
"DB:medium_format/name:095": "KiT Album",
"DB:medium_format/name:128": "DataPlay",
"DB:medium_format/name:129": "Mixed Mode CD",
"DB:medium_format/name:130": "DualDisc (DVD side)",
"DB:medium_format/name:131": "Betacam SP",
"DB:medium_format/name:164": "microSD",
"DB:medium_format/name:165": "Minimax CD",
"DB:medium_format/name:166": "MiniDVD",
"DB:medium_format/name:167": "Minimax DVD",
"DB:medium_format/name:168": "MiniDVD-Audio",
"DB:medium_format/name:169": "MiniDVD-Video",
"DB:medium_format/name:170": "Minimax DVD-Audio",
"DB:medium_format/name:171": "Minimax DVD-Video",
"DB:release_group_primary_type/name:001": "Album",
"DB:release_group_primary_type/name:002": "Single",
"DB:release_group_primary_type/name:003": "EP",
"DB:release_group_primary_type/name:011": "Other",
"DB:release_group_primary_type/name:012": "Broadcast",
"DB:release_group_secondary_type/name:001": "Compilation",
"DB:release_group_secondary_type/name:002": "Soundtrack",
"DB:release_group_secondary_type/name:003": "Spokenword",
"DB:release_group_secondary_type/name:004": "Interview",
"DB:release_group_secondary_type/name:005": "Audiobook",
"DB:release_group_secondary_type/name:006": "Live",
"DB:release_group_secondary_type/name:007": "Remix",
"DB:release_group_secondary_type/name:008": "DJ-mix",
"DB:release_group_secondary_type/name:009": "Mixtape/Street",
"DB:release_group_secondary_type/name:010": "Demo",
"DB:release_group_secondary_type/name:011": "Audio drama",
"DB:release_group_secondary_type/name:012": "Field recording",
"DB:release_status/name:001": "Official",
"DB:release_status/name:002": "Promotion",
"DB:release_status/name:003": "Bootleg",
"DB:release_status/name:004": "Pseudo-Release",
"DB:release_status/name:005": "Withdrawn",
"DB:release_status/name:006": "Cancelled",
}
|
routing | graph | from .track import TrackSend
try:
from sg_py_vendor.pymarshal import pm_assert
except ImportError:
from pymarshal import pm_assert
MAX_TRACK_SENDS = 16
class RoutingGraph:
def __init__(self, graph=None):
"""
@graph:
{int(track_num): {int(send_index): TrackSend}}
, where send_index is an arbitrary number starting from zero.
There may not be more than MAX_TRACK_SENDS track sends.
"""
self.graph = graph if graph is not None else {}
def reorder(self, a_dict):
"""
@a_dict:
{int: int}, A map of from index, to index
"""
self.graph = {a_dict[k]: v for k, v in self.graph.items()}
for k, f_dict in self.graph.items():
for v in f_dict.values():
v.track_num = k
v.output = a_dict[v.output]
def set_node(self, a_index, a_dict):
"""
a_index: int, the index to set
a_dict: {0: TrackSend(...), ...}
"""
self.graph[int(a_index)] = a_dict
def find_all_paths(self, start, end=0, path=[]):
path = path + [start]
if start == end:
return [path]
if not start in self.graph:
return []
paths = []
for node in (x.output for x in sorted(self.graph[start].values())):
if node not in path:
newpaths = self.find_all_paths(node, end, path)
for newpath in newpaths:
paths.append(newpath)
return paths
def check_for_feedback(self, a_new, a_old):
return self.find_all_paths(a_old, a_new)
def toggle(
self,
a_src,
a_dest,
conn_type=0,
):
"""
a_src: int, The track number of the source
a_dest: int, The track number of the destination
conn_type: int, 0: normal, 1 sidechain, 2: MIDI
"""
f_connected = a_src in self.graph and a_dest in [
x.output for x in self.graph[a_src].values() if x.conn_type == conn_type
]
if f_connected:
for k, v in self.graph[a_src].copy().items():
if v.output == a_dest and v.conn_type == conn_type:
self.graph[a_src].pop(k)
else:
if self.check_for_feedback(a_src, a_dest):
return "Can't make connection, it would create a feedback loop"
if a_src in self.graph and len(self.graph[a_src]) >= MAX_TRACK_SENDS:
return "All available sends already in use for " "track {}".format(
a_src
)
if not a_src in self.graph:
f_i = 0
self.graph[a_src] = {}
else:
for f_i in range(MAX_TRACK_SENDS):
if f_i not in self.graph[a_src]:
break
f_result = TrackSend(a_src, f_i, a_dest, conn_type)
self.graph[a_src][f_i] = f_result
self.set_node(a_src, self.graph[a_src])
return None
def set_default_output(
self,
a_track_num,
a_output=0,
):
"""Set a track to the default output (by default, track:0, main),
only if the track has no other connections.
@a_track_num: int, The track number to set a default output for
@a_output: int, The track number to connect @a_track_num to
"""
assert a_track_num != a_output
assert a_track_num != 0
if a_track_num not in self.graph or not self.graph[a_track_num]:
f_send = TrackSend(a_track_num, 0, a_output, 0)
self.set_node(a_track_num, {0: f_send})
return True
else:
return False
def sort_all_paths(self):
f_result = {}
for f_path in self.graph:
f_paths = self.find_all_paths(f_path, 0)
if f_paths:
f_result[f_path] = max(len(x) for x in f_paths)
else:
f_result[f_path] = 0
return sorted(
f_result,
key=lambda x: f_result[x],
reverse=True,
)
def __str__(self):
f_result = []
f_sorted = self.sort_all_paths()
f_result.append("|".join(str(x) for x in ("c", len(f_sorted))))
for f_index, f_i in zip(f_sorted, range(len(f_sorted))):
f_result.append("|".join(str(x) for x in ("t", f_index, f_i)))
for k in sorted(self.graph):
for v in sorted(self.graph[k].values()):
f_result.append(str(v))
f_result.append("\\")
return "\n".join(f_result)
@staticmethod
def from_str(a_str):
f_str = str(a_str)
f_result = RoutingGraph()
f_tracks = {}
for f_line in f_str.split("\n"):
if f_line == "\\":
break
f_line_arr = f_line.split("|")
f_uid = int(f_line_arr[1])
pm_assert(
f_line_arr[0] in ("t", "s", "c"),
IndexError,
)
if f_line_arr[0] == "t":
assert f_uid not in f_tracks
f_tracks[f_uid] = {}
elif f_line_arr[0] == "s":
f_send = TrackSend(*f_line_arr[1:])
f_tracks[f_uid][f_send.index] = f_send
elif f_line_arr[0] == "c":
pass
for k, v in f_tracks.items():
f_result.set_node(k, v)
return f_result
|
extractor | seeker | # coding: utf-8
from __future__ import unicode_literals
import re
from ..utils import get_element_by_class, strip_or_none
from .common import InfoExtractor
class SeekerIE(InfoExtractor):
_VALID_URL = (
r"https?://(?:www\.)?seeker\.com/(?P<display_id>.*)-(?P<article_id>\d+)\.html"
)
_TESTS = [
{
"url": "http://www.seeker.com/should-trump-be-required-to-release-his-tax-returns-1833805621.html",
"md5": "897d44bbe0d8986a2ead96de565a92db",
"info_dict": {
"id": "Elrn3gnY",
"ext": "mp4",
"title": "Should Trump Be Required To Release His Tax Returns?",
"description": "md5:41efa8cfa8d627841045eec7b018eb45",
"timestamp": 1490090165,
"upload_date": "20170321",
},
},
{
"url": "http://www.seeker.com/changes-expected-at-zoos-following-recent-gorilla-lion-shootings-1834116536.html",
"playlist": [
{
"md5": "0497b9f20495174be73ae136949707d2",
"info_dict": {
"id": "FihYQ8AE",
"ext": "mp4",
"title": "The Pros & Cons Of Zoos",
"description": "md5:d88f99a8ea8e7d25e6ff77f271b1271c",
"timestamp": 1490039133,
"upload_date": "20170320",
},
}
],
"info_dict": {
"id": "1834116536",
"title": "After Gorilla Killing, Changes Ahead for Zoos",
"description": "The largest association of zoos and others are hoping to learn from recent incidents that led to the shooting deaths of a gorilla and two lions.",
},
},
]
def _real_extract(self, url):
display_id, article_id = re.match(self._VALID_URL, url).groups()
webpage = self._download_webpage(url, display_id)
entries = []
for jwp_id in re.findall(r'data-video-id="([a-zA-Z0-9]{8})"', webpage):
entries.append(
self.url_result("jwplatform:" + jwp_id, "JWPlatform", jwp_id)
)
return self.playlist_result(
entries,
article_id,
self._og_search_title(webpage),
strip_or_none(get_element_by_class("subtitle__text", webpage))
or self._og_search_description(webpage),
)
|
decrypters | TnyCz | # -*- coding: utf-8 -*-
import re
from ..base.simple_decrypter import SimpleDecrypter
class TnyCz(SimpleDecrypter):
__name__ = "TnyCz"
__type__ = "decrypter"
__version__ = "0.09"
__status__ = "testing"
__pattern__ = r"http://(?:www\.)?tny\.cz/\w+"
__config__ = [
("enabled", "bool", "Activated", True),
("use_premium", "bool", "Use premium account if available", True),
(
"folder_per_package",
"Default;Yes;No",
"Create folder for each package",
"Default",
),
("max_wait", "int", "Reconnect if waiting time is greater than minutes", 10),
]
__description__ = """Tny.cz decrypter plugin"""
__license__ = "GPLv3"
__authors__ = [("Walter Purcaro", "vuolter@gmail.com")]
NAME_PATTERN = r"<title>(?P<N>.+?) - .+</title>"
def get_links(self):
m = re.search(r'<a id=\'save_paste\' href="(.+save\.php\?hash=.+)">', self.data)
return re.findall(".+", self.load(m.group(1))) if m else None
|
extractor | tdslifeway | from __future__ import unicode_literals
from .common import InfoExtractor
class TDSLifewayIE(InfoExtractor):
_VALID_URL = r"https?://tds\.lifeway\.com/v1/trainingdeliverysystem/courses/(?P<id>\d+)/index\.html"
_TEST = {
# From http://www.ministrygrid.com/training-viewer/-/training/t4g-2014-conference/the-gospel-by-numbers-4/the-gospel-by-numbers
"url": "http://tds.lifeway.com/v1/trainingdeliverysystem/courses/3453494717001/index.html?externalRegistration=AssetId%7C34F466F1-78F3-4619-B2AB-A8EFFA55E9E9%21InstanceId%7C0%21UserId%7Caaaaaaaa-aaaa-aaaa-aaaa-aaaaaaaaaaaa&grouping=http%3A%2F%2Flifeway.com%2Fvideo%2F3453494717001&activity_id=http%3A%2F%2Flifeway.com%2Fvideo%2F3453494717001&content_endpoint=http%3A%2F%2Ftds.lifeway.com%2Fv1%2Ftrainingdeliverysystem%2FScormEngineInterface%2FTCAPI%2Fcontent%2F&actor=%7B%22name%22%3A%5B%22Guest%20Guest%22%5D%2C%22account%22%3A%5B%7B%22accountServiceHomePage%22%3A%22http%3A%2F%2Fscorm.lifeway.com%2F%22%2C%22accountName%22%3A%22aaaaaaaa-aaaa-aaaa-aaaa-aaaaaaaaaaaa%22%7D%5D%2C%22objectType%22%3A%22Agent%22%7D&content_token=462a50b2-b6f9-4970-99b1-930882c499fb®istration=93d6ec8e-7f7b-4ed3-bbc8-a857913c0b2a&externalConfiguration=access%7CFREE%21adLength%7C-1%21assignOrgId%7C4AE36F78-299A-425D-91EF-E14A899B725F%21assignOrgParentId%7C%21courseId%7C%21isAnonymous%7Cfalse%21previewAsset%7Cfalse%21previewLength%7C-1%21previewMode%7Cfalse%21royalty%7CFREE%21sessionId%7C671422F9-8E79-48D4-9C2C-4EE6111EA1CD%21trackId%7C&auth=Basic%20OjhmZjk5MDBmLTBlYTMtNDJhYS04YjFlLWE4MWQ3NGNkOGRjYw%3D%3D&endpoint=http%3A%2F%2Ftds.lifeway.com%2Fv1%2Ftrainingdeliverysystem%2FScormEngineInterface%2FTCAPI%2F",
"info_dict": {
"id": "3453494717001",
"ext": "mp4",
"title": "The Gospel by Numbers",
"thumbnail": r"re:^https?://.*\.jpg",
"upload_date": "20140410",
"description": "Coming soon from T4G 2014!",
"uploader_id": "2034960640001",
"timestamp": 1397145591,
},
"params": {
# m3u8 download
"skip_download": True,
},
"add_ie": ["BrightcoveNew"],
}
BRIGHTCOVE_URL_TEMPLATE = "http://players.brightcove.net/2034960640001/default_default/index.html?videoId=%s"
def _real_extract(self, url):
brightcove_id = self._match_id(url)
return self.url_result(
self.BRIGHTCOVE_URL_TEMPLATE % brightcove_id, "BrightcoveNew", brightcove_id
)
|
torrent-checker | torrent_checker | import asyncio
import logging
import random
import time
from asyncio import CancelledError
from collections import defaultdict
from typing import Dict, List, Optional, Tuple, Union
from ipv8.taskmanager import TaskManager
from pony.orm import db_session, desc, select
from pony.utils import between
from tribler.core import notifications
from tribler.core.components.libtorrent.download_manager.download_manager import (
DownloadManager,
)
from tribler.core.components.metadata_store.db.serialization import REGULAR_TORRENT
from tribler.core.components.metadata_store.db.store import MetadataStore
from tribler.core.components.torrent_checker.torrent_checker import DHT
from tribler.core.components.torrent_checker.torrent_checker.dataclasses import (
HEALTH_FRESHNESS_SECONDS,
HealthInfo,
TrackerResponse,
)
from tribler.core.components.torrent_checker.torrent_checker.torrentchecker_session import (
FakeBep33DHTSession,
FakeDHTSession,
TrackerSession,
UdpSocketManager,
create_tracker_session,
)
from tribler.core.components.torrent_checker.torrent_checker.tracker_manager import (
MAX_TRACKER_FAILURES,
TrackerManager,
)
from tribler.core.components.torrent_checker.torrent_checker.utils import (
aggregate_responses_for_infohash,
filter_non_exceptions,
gather_coros,
)
from tribler.core.config.tribler_config import TriblerConfig
from tribler.core.utilities.notifier import Notifier
from tribler.core.utilities.tracker_utils import MalformedTrackerURLException
from tribler.core.utilities.unicode import hexlify
from tribler.core.utilities.utilities import has_bep33_support, is_valid_url
TRACKER_SELECTION_INTERVAL = 1 # The interval for querying a random tracker
TORRENT_SELECTION_INTERVAL = (
120 # The interval for checking the health of a random torrent
)
USER_CHANNEL_TORRENT_SELECTION_INTERVAL = (
10 * 60
) # The interval for checking the health of torrents in user's channel.
MIN_TORRENT_CHECK_INTERVAL = (
900 # How much time we should wait before checking a torrent again
)
TORRENT_CHECK_RETRY_INTERVAL = (
30 # Interval when the torrent was successfully checked for the last time
)
MAX_TORRENTS_CHECKED_PER_SESSION = 50
TORRENT_SELECTION_POOL_SIZE = (
2 # How many torrents to check (popular or random) during periodic check
)
USER_CHANNEL_TORRENT_SELECTION_POOL_SIZE = (
5 # How many torrents to check from user's channel during periodic check
)
TORRENTS_CHECKED_RETURN_SIZE = (
240 # Estimated torrents checked on default 4 hours idle run
)
class TorrentChecker(TaskManager):
def __init__(
self,
config: TriblerConfig,
download_manager: DownloadManager,
notifier: Notifier,
tracker_manager: TrackerManager,
metadata_store: MetadataStore,
socks_listen_ports: Optional[List[int]] = None,
):
super().__init__()
self._logger = logging.getLogger(self.__class__.__name__)
self.tracker_manager = tracker_manager
self.mds = metadata_store
self.download_manager = download_manager
self.notifier = notifier
self.config = config
self.socks_listen_ports = socks_listen_ports
self._should_stop = False
self._sessions = defaultdict(list)
self.socket_mgr = UdpSocketManager()
self.udp_transport = None
# We keep track of the results of popular torrents checked by you.
# The popularity community gossips this information around.
self._torrents_checked: Optional[Dict[bytes, HealthInfo]] = None
async def initialize(self):
self.register_task(
"check random tracker",
self.check_random_tracker,
interval=TRACKER_SELECTION_INTERVAL,
)
self.register_task(
"check local torrents",
self.check_local_torrents,
interval=TORRENT_SELECTION_INTERVAL,
)
self.register_task(
"check channel torrents",
self.check_torrents_in_user_channel,
interval=USER_CHANNEL_TORRENT_SELECTION_INTERVAL,
)
await self.create_socket_or_schedule()
async def listen_on_udp(self):
loop = asyncio.get_event_loop()
transport, _ = await loop.create_datagram_endpoint(
lambda: self.socket_mgr, local_addr=("0.0.0.0", 0)
)
return transport
async def create_socket_or_schedule(self):
"""
This method attempts to bind to a UDP port. If it fails for some reason (i.e. no network connection), we try
again later.
"""
try:
self.udp_transport = await self.listen_on_udp()
except OSError as e:
self._logger.error(
"Error when creating UDP socket in torrent checker: %s", e
)
self.register_task(
"listen_udp_port", self.create_socket_or_schedule, delay=10
)
async def shutdown(self):
"""
Shutdown the torrent health checker.
Once shut down it can't be started again.
:returns A deferred that will fire once the shutdown has completed.
"""
self._should_stop = True
if self.udp_transport:
self.udp_transport.close()
self.udp_transport = None
await self.shutdown_task_manager()
async def check_random_tracker(self):
"""
Calling this method will fetch a random tracker from the database, select some torrents that have this
tracker, and perform a request to these trackers.
Return whether the check was successful.
"""
if self._should_stop:
self._logger.warning(
"Not performing tracker check since we are shutting down"
)
return
tracker = self.get_next_tracker()
if not tracker:
self._logger.warning(
"No tracker to select from to check torrent health, skip"
)
return
# get the torrents that should be checked
url = tracker.url
with db_session:
dynamic_interval = TORRENT_CHECK_RETRY_INTERVAL * (2**tracker.failures)
torrents = select(
ts
for ts in tracker.torrents
if ts.last_check + dynamic_interval < int(time.time())
)
infohashes = [
t.infohash for t in torrents[:MAX_TORRENTS_CHECKED_PER_SESSION]
]
if len(infohashes) == 0:
# We have no torrent to recheck for this tracker. Still update the last_check for this tracker.
self._logger.info(f"No torrent to check for tracker {url}")
self.tracker_manager.update_tracker_info(url)
return
try:
session = self._create_session_for_request(url, timeout=30)
except MalformedTrackerURLException as e:
session = None
# Remove the tracker from the database
self.tracker_manager.remove_tracker(url)
self._logger.warning(e)
if session is None:
self._logger.warning(
"A session cannot be created. The torrent check procedure has been cancelled."
)
return
# We shuffle the list so that different infohashes are checked on subsequent scrape requests if the total
# number of infohashes exceeds the maximum number of infohashes we check.
random.shuffle(infohashes)
for infohash in infohashes:
session.add_infohash(infohash)
self._logger.info(
f"Selected {len(infohashes)} new torrents to check on random tracker: {url}"
)
try:
response = await self.get_tracker_response(session)
except Exception as e: # pylint: disable=broad-except
self._logger.warning(e)
else:
health_list = response.torrent_health_list
self._logger.info(
f"Received {len(health_list)} health info results from tracker: {health_list}"
)
async def get_tracker_response(self, session: TrackerSession) -> TrackerResponse:
t1 = time.time()
try:
result = await session.connect_to_tracker()
except CancelledError:
self._logger.info(
f"Tracker session is being cancelled: {session.tracker_url}"
)
raise
except Exception as e:
exception_str = str(e).replace("\n]", "]")
self._logger.warning(
f"Got session error for the tracker: {session.tracker_url}\n{exception_str}"
)
self.tracker_manager.update_tracker_info(session.tracker_url, False)
raise e
finally:
await self.clean_session(session)
t2 = time.time()
self._logger.info(
f"Got response from {session.__class__.__name__} in {t2 - t1:.3f} seconds: {result}"
)
with db_session:
for health in result.torrent_health_list:
self.update_torrent_health(health)
return result
@property
def torrents_checked(self) -> Dict[bytes, HealthInfo]:
if self._torrents_checked is None:
self._torrents_checked = self.load_torrents_checked_from_db()
lines = "\n".join(
f" {health}"
for health in sorted(
self._torrents_checked.values(),
key=lambda health: -health.last_check,
)
)
self._logger.info(f"Initially loaded self-checked torrents:\n{lines}")
return self._torrents_checked
@db_session
def load_torrents_checked_from_db(self) -> Dict[bytes, HealthInfo]:
result = {}
now = int(time.time())
last_fresh_time = now - HEALTH_FRESHNESS_SECONDS
checked_torrents = list(
self.mds.TorrentState.select(
lambda g: g.has_data
and g.self_checked
and between(g.last_check, last_fresh_time, now)
)
.order_by(lambda g: (desc(g.seeders), g.last_check))
.limit(TORRENTS_CHECKED_RETURN_SIZE)
)
for torrent in checked_torrents:
result[torrent.infohash] = HealthInfo(
torrent.infohash,
torrent.seeders,
torrent.leechers,
last_check=torrent.last_check,
self_checked=True,
)
return result
@db_session
def torrents_to_check(self):
"""
Two categories of torrents are selected (popular & old). From the pool of selected torrents, a certain
number of them are submitted for health check. The torrents that are within the freshness window are
excluded from the selection considering the health information is still fresh.
1. Popular torrents (50%)
The indicator for popularity here is considered as the seeder count with direct proportionality
assuming more seeders -> more popular. There could be other indicators to be introduced later.
2. Old torrents (50%)
By old torrents, we refer to those checked quite farther in the past, sorted by the last_check value.
"""
last_fresh_time = time.time() - HEALTH_FRESHNESS_SECONDS
popular_torrents = list(
self.mds.TorrentState.select(lambda g: g.last_check < last_fresh_time)
.order_by(lambda g: (desc(g.seeders), g.last_check))
.limit(TORRENT_SELECTION_POOL_SIZE)
)
old_torrents = list(
self.mds.TorrentState.select(lambda g: g.last_check < last_fresh_time)
.order_by(lambda g: (g.last_check, desc(g.seeders)))
.limit(TORRENT_SELECTION_POOL_SIZE)
)
selected_torrents = popular_torrents + old_torrents
selected_torrents = random.sample(
selected_torrents, min(TORRENT_SELECTION_POOL_SIZE, len(selected_torrents))
)
return selected_torrents
async def check_local_torrents(self) -> Tuple[List, List]:
"""
Perform a full health check on a few popular and old torrents in the database.
"""
selected_torrents = self.torrents_to_check()
self._logger.info(f"Check {len(selected_torrents)} local torrents")
coros = [self.check_torrent_health(t.infohash) for t in selected_torrents]
results = await gather_coros(coros)
self._logger.info(f"Results for local torrents check: {results}")
return selected_torrents, results
@db_session
def torrents_to_check_in_user_channel(self):
"""
Returns a list of outdated torrents of user's channel which
has not been checked recently.
"""
last_fresh_time = time.time() - HEALTH_FRESHNESS_SECONDS
channel_torrents = list(
self.mds.TorrentMetadata.select(
lambda g: g.public_key == self.mds.my_public_key_bin
and g.metadata_type == REGULAR_TORRENT
and g.health.last_check < last_fresh_time
)
.order_by(lambda g: g.health.last_check)
.limit(USER_CHANNEL_TORRENT_SELECTION_POOL_SIZE)
)
return channel_torrents
async def check_torrents_in_user_channel(
self
) -> List[Union[HealthInfo, BaseException]]:
"""
Perform a full health check of torrents in user's channel
"""
selected_torrents = self.torrents_to_check_in_user_channel()
self._logger.info(f"Check {len(selected_torrents)} torrents in user channel")
coros = [self.check_torrent_health(t.infohash) for t in selected_torrents]
results = await gather_coros(coros)
self._logger.info(f"Results for torrents in user channel: {results}")
return results
def get_next_tracker(self):
while tracker := self.tracker_manager.get_next_tracker():
url = tracker.url
if not is_valid_url(url):
self.tracker_manager.remove_tracker(url)
elif tracker.failures >= MAX_TRACKER_FAILURES:
self.tracker_manager.update_tracker_info(url, is_successful=False)
else:
return tracker
return None
def is_blacklisted_tracker(self, tracker_url):
return tracker_url in self.tracker_manager.blacklist
@db_session
def get_valid_trackers_of_torrent(self, infohash):
"""Get a set of valid trackers for torrent. Also remove any invalid torrent."""
db_tracker_list = self.mds.TorrentState.get(infohash=infohash).trackers
return {
tracker.url
for tracker in db_tracker_list
if is_valid_url(tracker.url)
and not self.is_blacklisted_tracker(tracker.url)
}
async def check_torrent_health(
self, infohash: bytes, timeout=20, scrape_now=False
) -> HealthInfo:
"""
Check the health of a torrent with a given infohash.
:param infohash: Torrent infohash.
:param timeout: The timeout to use in the performed requests
:param scrape_now: Flag whether we want to force scraping immediately
"""
infohash_hex = hexlify(infohash)
self._logger.info(f"Check health for the torrent: {infohash_hex}")
tracker_set = []
# We first check whether the torrent is already in the database and checked before
with db_session:
torrent_state = self.mds.TorrentState.get(infohash=infohash)
if torrent_state:
last_check = torrent_state.last_check
time_diff = time.time() - last_check
if time_diff < MIN_TORRENT_CHECK_INTERVAL and not scrape_now:
self._logger.info(
f"Time interval too short, not doing torrent health check for {infohash_hex}"
)
return torrent_state.to_health()
# get torrent's tracker list from DB
tracker_set = self.get_valid_trackers_of_torrent(torrent_state.infohash)
self._logger.info(f"Trackers for {infohash_hex}: {tracker_set}")
coros = []
for tracker_url in tracker_set:
if session := self._create_session_for_request(
tracker_url, timeout=timeout
):
session.add_infohash(infohash)
coros.append(self.get_tracker_response(session))
session_cls = FakeBep33DHTSession if has_bep33_support() else FakeDHTSession
session = session_cls(self.download_manager, timeout)
session.add_infohash(infohash)
self._logger.info(f"DHT session has been created for {infohash_hex}: {session}")
self._sessions[DHT].append(session)
coros.append(self.get_tracker_response(session))
responses = await gather_coros(coros)
self._logger.info(
f"{len(responses)} responses for {infohash_hex} have been received: {responses}"
)
successful_responses = filter_non_exceptions(responses)
health = aggregate_responses_for_infohash(infohash, successful_responses)
if (
health.last_check == 0
): # if not zero, was already updated in get_tracker_response
health.last_check = int(time.time())
health.self_checked = True
self.update_torrent_health(health)
def _create_session_for_request(
self, tracker_url, timeout=20
) -> Optional[TrackerSession]:
self._logger.debug(f"Creating a session for the request: {tracker_url}")
required_hops = self.config.download_defaults.number_hops
actual_hops = len(self.socks_listen_ports or [])
if required_hops > actual_hops:
self._logger.warning(
f"Dropping the request. Required amount of hops doesn't reached. "
f"Required hops: {required_hops}. Actual hops: {actual_hops}"
)
return None
proxy = (
("127.0.0.1", self.socks_listen_ports[required_hops - 1])
if required_hops > 0
else None
)
session = create_tracker_session(tracker_url, timeout, proxy, self.socket_mgr)
self._logger.info(f"Tracker session has been created: {session}")
self._sessions[tracker_url].append(session)
return session
async def clean_session(self, session):
url = session.tracker_url
self.tracker_manager.update_tracker_info(url, not session.is_failed)
# Remove the session from our session list dictionary
self._sessions[url].remove(session)
if len(self._sessions[url]) == 0 and url != DHT:
del self._sessions[url]
await session.cleanup()
self._logger.debug("Session has been cleaned up")
def update_torrent_health(self, health: HealthInfo) -> bool:
"""
Updates the torrent state in the database if it already exists, otherwise do nothing.
Returns True if the update was successful, False otherwise.
"""
if not health.is_valid():
self._logger.warning(f"Invalid health info ignored: {health}")
return False
if not health.self_checked:
self._logger.error(f"Self-checked torrent health expected. Got: {health}")
return False
self._logger.debug(f"Update torrent health: {health}")
with db_session:
# Update torrent state
torrent_state = self.mds.TorrentState.get_for_update(
infohash=health.infohash
)
if not torrent_state:
self._logger.warning(f"Unknown torrent: {hexlify(health.infohash)}")
return False
prev_health = torrent_state.to_health()
if not health.should_replace(prev_health):
self._logger.info(
"Skip health update, the health in the database is fresher or have more seeders"
)
self.notify(prev_health) # to update UI state from "Checking..."
return False
torrent_state.set(
seeders=health.seeders,
leechers=health.leechers,
last_check=health.last_check,
self_checked=True,
)
if health.seeders > 0 or health.leechers > 0:
self.torrents_checked[health.infohash] = health
else:
self.torrents_checked.pop(health.infohash, None)
self.notify(health)
return True
def notify(self, health: HealthInfo):
self.notifier[notifications.channel_entity_updated](
{
"infohash": health.infohash_hex,
"num_seeders": health.seeders,
"num_leechers": health.leechers,
"last_tracker_check": health.last_check,
"health": "updated",
}
)
|
plugins | okru | """
$description Russian live-streaming and video hosting social platform.
$url ok.ru
$url mobile.ok.ru
$type live, vod
$metadata id
$metadata author
$metadata title
"""
import logging
import re
from urllib.parse import unquote, urlparse
from streamlink.plugin import Plugin, pluginmatcher
from streamlink.plugin.api import validate
from streamlink.stream.dash import DASHStream
from streamlink.stream.hls import HLSStream
from streamlink.stream.http import HTTPStream
log = logging.getLogger(__name__)
@pluginmatcher(re.compile(r"https?://(?:www\.)?ok\.ru/"))
@pluginmatcher(re.compile(r"https?://m(?:obile)?\.ok\.ru/"))
class OKru(Plugin):
QUALITY_WEIGHTS = {
"full": 1080,
"1080": 1080,
"hd": 720,
"720": 720,
"sd": 480,
"480": 480,
"360": 360,
"low": 360,
"lowest": 240,
"mobile": 144,
}
@classmethod
def stream_weight(cls, key):
weight = cls.QUALITY_WEIGHTS.get(key)
if weight:
return weight, "okru"
return super().stream_weight(key)
def _get_streams_mobile(self):
data = self.session.http.get(
self.url,
schema=validate.Schema(
validate.parse_html(),
validate.xml_find(".//a[@data-video]"),
validate.get("data-video"),
validate.none_or_all(
str,
validate.parse_json(),
{
"videoName": str,
"videoSrc": validate.url(),
"movieId": str,
},
validate.union_get("movieId", "videoName", "videoSrc"),
),
),
)
if not data:
return
self.id, self.title, url = data
stream_url = self.session.http.head(url).headers.get("Location")
if not stream_url:
return
return (
HLSStream.parse_variant_playlist(self.session, stream_url)
if urlparse(stream_url).path.endswith(".m3u8")
else {"vod": HTTPStream(self.session, stream_url)}
)
def _get_streams_default(self):
schema_metadata = validate.Schema(
validate.parse_json(),
{
validate.optional("author"): validate.all({"name": str}, validate.get("name")),
validate.optional("movie"): validate.all({"title": str}, validate.get("title")),
validate.optional("hlsManifestUrl"): validate.url(),
validate.optional("hlsMasterPlaylistUrl"): validate.url(),
validate.optional("liveDashManifestUrl"): validate.url(),
validate.optional("videos"): [
validate.all(
{
"name": str,
"url": validate.url(),
},
validate.union_get("name", "url"),
)
],
},
)
metadata, metadata_url = self.session.http.get(
self.url,
schema=validate.Schema(
validate.parse_html(),
validate.xml_find(".//*[@data-options]"),
validate.get("data-options"),
validate.parse_json(),
{
"flashvars": {
validate.optional("metadata"): str,
validate.optional("metadataUrl"): validate.all(
validate.transform(unquote),
validate.url(),
),
}
},
validate.get("flashvars"),
validate.union_get("metadata", "metadataUrl"),
),
)
self.session.http.headers.update({"Referer": self.url})
if not metadata and metadata_url:
metadata = self.session.http.post(metadata_url).text
log.trace(f"{metadata!r}")
data = schema_metadata.validate(metadata)
self.author = data.get("author")
self.title = data.get("movie")
for hls_url in data.get("hlsManifestUrl"), data.get("hlsMasterPlaylistUrl"):
if hls_url is not None:
return HLSStream.parse_variant_playlist(self.session, hls_url)
if data.get("liveDashManifestUrl"):
return DASHStream.parse_manifest(self.session, data.get("liveDashManifestUrl"))
return {
f"{self.QUALITY_WEIGHTS[name]}p" if name in self.QUALITY_WEIGHTS else name: HTTPStream(self.session, url)
for name, url in data.get("videos", [])
}
def _get_streams(self):
return self._get_streams_default() if self.matches[0] else self._get_streams_mobile()
__plugin__ = OKru
|
Draft | TestDraftGui | # ***************************************************************************
# * Copyright (c) 2013 Yorik van Havre <yorik@uncreated.net> *
# * Copyright (c) 2020 Eliud Cabrera Castillo <e.cabrera-castillo@tum.de> *
# * *
# * This file is part of the FreeCAD CAx development system. *
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU Lesser General Public License (LGPL) *
# * as published by the Free Software Foundation; either version 2 of *
# * the License, or (at your option) any later version. *
# * for detail see the LICENCE text file. *
# * *
# * FreeCAD is distributed in the hope that it will be useful, *
# * but WITHOUT ANY WARRANTY; without even the implied warranty of *
# * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
# * GNU Library General Public License for more details. *
# * *
# * You should have received a copy of the GNU Library General Public *
# * License along with FreeCAD; if not, write to the Free Software *
# * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *
# * USA *
# * *
# ***************************************************************************
"""Unit tests for the Draft workbench, GUI only.
From the terminal, run the following:
FreeCAD -t TestDraftGui
From within FreeCAD, run the following:
import Test, TestDraftGui
Test.runTestsFromModule(TestDraftGui)
For the non-GUI tests see TestDraft.
"""
# ===========================================================================
# The unit tests can be run from the operating system terminal, or from
# within FreeCAD itself.
#
# The tests can be run using the full 'FreeCAD' executable
# or the console only 'FreeCADCmd' executable. In the latter case
# some functions cannot be tested as the view providers (visual properties)
# are not available.
#
# ===========================================================================
# In the following, first the command to run the test from the operating
# system terminal is listed, followed by the commands to run the test
# from the Python console within FreeCAD.
#
# ===========================================================================
# Run all Draft tests
# ----
# FreeCAD -t TestDraft
#
# >>> import Test, TestDraft
# >>> Test.runTestsFromModule(TestDraft)
#
# ===========================================================================
# Run tests from a specific module (all classes within this module)
# ----
# FreeCAD -t drafttests.test_creation
#
# >>> import Test, drafttests.test_creation
# >>> Test.runTestsFromModule(drafttests.test_creation)
#
# ===========================================================================
# Run tests from a specific class within a module
# ----
# FreeCAD -t drafttests.test_creation.DraftCreation
#
# >>> import Test, drafttests.test_creation
# >>> Test.runTestsFromClass(drafttests.test_creation.DraftCreation)
#
# ===========================================================================
# Run a specific unit test from a class within a module
# ----
# FreeCAD -t drafttests.test_creation.DraftCreation.test_line
#
# >>> import unittest
# >>> one_test = "drafttests.test_creation.DraftCreation.test_line"
# >>> all_tests = unittest.TestLoader().loadTestsFromName(one_test)
# >>> unittest.TextTestRunner().run(all_tests)
# ===========================================================================
# When the full test is run
# FreeCAD -t TestDraft
#
# all classes that are found in this file are run.
#
# We import the classes from submodules. These classes contain
# the actual unit tests.
#
# The classes will be run in alphabetical order. So, to force
# a particular order of testing we import them with a name
# that follows a defined alphanumeric sequence.
# Import tests
from drafttests.test_import_gui import DraftGuiImport as DraftTestGui01
from drafttests.test_import_tools import DraftImportTools as DraftTestGui02
from drafttests.test_pivy import DraftPivy as DraftTestGui03
# Use the modules so that code checkers don't complain (flake8)
True if DraftTestGui01 else False
True if DraftTestGui02 else False
True if DraftTestGui03 else False
|
accounts | DebridlinkFr | # -*- coding: utf-8 -*-
import json
import time
import pycurl
from pyload.core.network.http.exceptions import BadHeader
from ..base.multi_account import MultiAccount
from ..downloaders.DebridlinkFr import error_description
class DebridlinkFr(MultiAccount):
__name__ = "DebridlinkFr"
__type__ = "account"
__version__ = "0.06"
__status__ = "testing"
__config__ = [
("mh_mode", "all;listed;unlisted", "Filter downloaders to use", "all"),
("mh_list", "str", "Downloader list (comma separated)", ""),
("mh_interval", "int", "Reload interval in hours", 12),
]
__description__ = """Debridlink.fr account plugin"""
__license__ = "GPLv3"
__authors__ = [("GammaC0de", "nitzo2001[AT]yahoo[DOT]com")]
TUNE_TIMEOUT = False
#: See https://debrid-link.fr/api_doc/v2
API_URL = "https://debrid-link.fr/api/"
def api_request(self, method, get={}, post={}):
api_token = self.info["data"].get("api_token", None)
if api_token and method != "oauth/token":
self.req.http.c.setopt(
pycurl.HTTPHEADER, ["Authorization: Bearer " + api_token]
)
self.req.http.c.setopt(
pycurl.USERAGENT, "pyLoad/{}".format(self.pyload.version)
)
try:
json_data = self.load(self.API_URL + method, get=get, post=post)
except BadHeader as exc:
json_data = exc.content
return json.loads(json_data)
def _refresh_token(self, client_id, refresh_token):
api_data = self.api_request(
"oauth/token",
post={
"client_id": client_id,
"refresh_token": refresh_token,
"grant_type": "refresh_token",
},
)
if "error" in api_data:
if api_data["error"] == "invalid_request":
self.log_error(
self._(
"You have to use GetDebridlinkToken.py to authorize pyLoad: "
"https://github.com/pyload/pyload/files/9353788/GetDebridlinkToken.zip"
)
)
else:
self.log_error(
api_data.get(
"error_description", error_description(api_data["error"])
)
)
self.fail_login()
return api_data["access_token"], api_data["expires_in"]
def grab_hosters(self, user, password, data):
api_data = self.api_request("v2/downloader/hostnames")
if api_data["success"]:
return api_data["value"]
else:
return []
def grab_info(self, user, password, data):
api_data = self.api_request("v2/account/infos")
if api_data["success"]:
premium = api_data["value"]["premiumLeft"] > 0
validuntil = api_data["value"]["premiumLeft"] + time.time()
else:
self.log_error(
self._("Unable to retrieve account information"),
api_data.get("error_description", error_description(api_data["error"])),
)
validuntil = None
premium = None
return {"validuntil": validuntil, "trafficleft": -1, "premium": premium}
def signin(self, user, password, data):
if "token" not in data:
api_token, timeout = self._refresh_token(user, password)
data["api_token"] = api_token
self.timeout = timeout - 5 * 60 #: Five minutes less to be on the safe side
api_data = self.api_request("v2/account/infos")
if "error" in api_data:
if api_data["error"] == "badToken": #: Token expired? try to refresh
api_token, timeout = self._refresh_token(user, password)
data["api_token"] = api_token
self.timeout = (
timeout - 5 * 60
) #: Five minutes less to be on the safe side
else:
self.log_error(
api_data.get(
"error_description", error_description(api_data["error"])
)
)
self.fail_login()
|
queries | session_recording_list_from_replay_summary | import dataclasses
import re
from datetime import datetime, timedelta
from typing import Any, Dict, List, Literal, NamedTuple, Tuple, Union
from django.conf import settings
from posthog.client import sync_execute
from posthog.cloud_utils import is_cloud
from posthog.constants import (
TREND_FILTER_TYPE_ACTIONS,
AvailableFeature,
PropertyOperatorType,
)
from posthog.models import Entity
from posthog.models.action.util import format_entity_filter
from posthog.models.filters.mixins.utils import cached_property
from posthog.models.filters.session_recordings_filter import SessionRecordingsFilter
from posthog.models.instance_setting import get_instance_setting
from posthog.models.property import PropertyGroup
from posthog.models.property.util import parse_prop_grouped_clauses
from posthog.models.team import PersonOnEventsMode
from posthog.models.team.team import Team
from posthog.queries.event_query import EventQuery
from posthog.queries.util import PersonPropertiesMode
@dataclasses.dataclass(frozen=True)
class SummaryEventFiltersSQL:
having_conditions: str
having_select: str
where_conditions: str
params: Dict[str, Any]
class SessionRecordingQueryResult(NamedTuple):
results: List
has_more_recording: bool
def _get_recording_start_time_clause(
recording_filters: SessionRecordingsFilter
) -> Tuple[str, Dict[str, Any]]:
start_time_clause = ""
start_time_params = {}
if recording_filters.date_from:
start_time_clause += "\nAND start_time >= %(start_time)s"
start_time_params["start_time"] = recording_filters.date_from
if recording_filters.date_to:
start_time_clause += "\nAND start_time <= %(end_time)s"
start_time_params["end_time"] = recording_filters.date_to
return start_time_clause, start_time_params
def _get_filter_by_provided_session_ids_clause(
recording_filters: SessionRecordingsFilter, column_name="session_id"
) -> Tuple[str, Dict[str, Any]]:
if recording_filters.session_ids is None:
return "", {}
return f'AND "{column_name}" in %(session_ids)s', {
"session_ids": recording_filters.session_ids
}
def ttl_days(team: Team) -> int:
ttl_days = (get_instance_setting("RECORDINGS_TTL_WEEKS") or 3) * 7
if is_cloud():
# NOTE: We use Playlists as a proxy to see if they are subbed to Recordings
is_paid = team.organization.is_feature_available(
AvailableFeature.RECORDINGS_PLAYLISTS
)
ttl_days = (
settings.REPLAY_RETENTION_DAYS_MAX
if is_paid
else settings.REPLAY_RETENTION_DAYS_MIN
)
# NOTE: The date we started reliably ingested data to blob storage
days_since_blob_ingestion = (datetime.now() - datetime(2023, 8, 1)).days
if days_since_blob_ingestion < ttl_days:
ttl_days = days_since_blob_ingestion
return ttl_days
class PersonsQuery(EventQuery):
_filter: SessionRecordingsFilter
# we have to implement this from EventQuery but don't need it
def _determine_should_join_distinct_ids(self) -> None:
pass
# we have to implement this from EventQuery but don't need it
def _data_to_return(self, results: List[Any]) -> List[Dict[str, Any]]:
pass
_raw_persons_query = """
SELECT distinct_id, argMax(person_id, version) as person_id
{select_person_props}
FROM person_distinct_id2 as pdi
{filter_persons_clause}
WHERE team_id = %(team_id)s
{prop_filter_clause}
GROUP BY distinct_id
HAVING
argMax(is_deleted, version) = 0
{prop_having_clause}
{filter_by_person_uuid_condition}
"""
def get_query(self) -> Tuple[str, Dict[str, Any]]:
prop_query, prop_params = self._get_prop_groups(
PropertyGroup(
type=PropertyOperatorType.AND,
values=[
g
for g in self._filter.property_groups.flat
if g.type == "person" or "cohort" in g.type
],
),
person_id_joined_alias=f"{self.DISTINCT_ID_TABLE_ALIAS}.person_id",
)
# hogql person props queries rely on an aggregated column and so have to go in the having clause
# not the where clause
having_prop_query, having_prop_params = self._get_prop_groups(
PropertyGroup(
type=PropertyOperatorType.AND,
values=[
g
for g in self._filter.property_groups.flat
if g.type == "hogql" and "person.properties" in g.key
],
),
person_id_joined_alias=f"{self.DISTINCT_ID_TABLE_ALIAS}.person_id",
)
person_query, person_query_params = self._get_person_query()
should_join_persons = self._filter.person_uuid or person_query
if not should_join_persons:
return "", {}
else:
filter_persons_clause = person_query or ""
filter_by_person_uuid_condition = (
"and person_id = %(person_uuid)s" if self._filter.person_uuid else ""
)
return self._raw_persons_query.format(
filter_persons_clause=filter_persons_clause,
select_person_props=", argMax(person_props, version) as person_props"
if "person_props" in filter_persons_clause
else "",
prop_filter_clause=prop_query,
prop_having_clause=having_prop_query,
filter_by_person_uuid_condition=filter_by_person_uuid_condition,
), {
"team_id": self._team_id,
**person_query_params,
"person_uuid": self._filter.person_uuid,
**prop_params,
**having_prop_params,
}
class SessionIdEventsQuery(EventQuery):
_filter: SessionRecordingsFilter
# we have to implement this from EventQuery but don't need it
def _determine_should_join_distinct_ids(self) -> None:
pass
# we have to implement this from EventQuery but don't need it
def _data_to_return(self, results: List[Any]) -> List[Dict[str, Any]]:
pass
def _determine_should_join_events(self):
filters_by_event_or_action = (
self._filter.entities and len(self._filter.entities) > 0
)
# for e.g. test account filters might have event properties without having an event or action filter
has_event_property_filters = (
len(
[
pg
for pg in self._filter.property_groups.flat
# match when it is an event property filter
# or if its hogql and the key contains "properties." but not "person.properties."
# it's ok to match if there's both "properties." and "person.properties." in the key
# but not when its only "person.properties."
if pg.type == "event"
or pg.type == "hogql"
and re.search(r"(?<!person\.)properties\.", pg.key)
]
)
> 0
)
return filters_by_event_or_action or has_event_property_filters
def __init__(
self,
**kwargs,
):
super().__init__(
**kwargs,
)
@property
def ttl_days(self):
return ttl_days(self._team)
_raw_events_query = """
SELECT
{select_event_ids}
{event_filter_having_events_select}
`$session_id`
FROM events e
-- sometimes we have to join on persons so we can access e.g. person_props in filters
{persons_join}
PREWHERE
team_id = %(team_id)s
-- regardless of what other filters are applied
-- limit by storage TTL
AND e.timestamp >= %(clamped_to_storage_ttl)s
AND e.timestamp <= now()
WHERE
notEmpty(`$session_id`)
{events_timestamp_clause}
{event_filter_where_conditions}
{prop_filter_clause}
{provided_session_ids_clause}
-- other times we can check distinct id against a sub query which should be faster than joining
{persons_sub_query}
GROUP BY `$session_id`
HAVING 1=1 {event_filter_having_events_condition}
"""
def format_event_filter(
self, entity: Entity, prepend: str, team_id: int
) -> Tuple[str, Dict[str, Any]]:
filter_sql, params = format_entity_filter(
team_id=team_id,
entity=entity,
prepend=prepend,
filter_by_team=False,
person_id_joined_alias=f"{self.DISTINCT_ID_TABLE_ALIAS}.person_id",
hogql_context=self._filter.hogql_context,
)
filters, filter_params = parse_prop_grouped_clauses(
team_id=team_id,
property_group=entity.property_groups,
prepend=prepend,
allow_denormalized_props=True,
has_person_id_joined=True,
person_properties_mode=PersonPropertiesMode.USING_PERSON_PROPERTIES_COLUMN,
hogql_context=self._filter.hogql_context,
)
filter_sql += f" {filters}"
params = {**params, **filter_params}
return filter_sql, params
@cached_property
def build_event_filters(self) -> SummaryEventFiltersSQL:
event_names_to_filter: List[Union[int, str]] = []
params: Dict = {}
condition_sql = ""
for index, entity in enumerate(self._filter.entities):
if entity.type == TREND_FILTER_TYPE_ACTIONS:
action = entity.get_action()
event_names_to_filter.extend(
[
ae
for ae in action.get_step_events()
if ae not in event_names_to_filter
]
)
else:
if entity.id and entity.id not in event_names_to_filter:
event_names_to_filter.append(entity.id)
(
this_entity_condition_sql,
this_entity_filter_params,
) = self.format_event_filter(
entity, prepend=f"event_matcher_{index}", team_id=self._team_id
)
joining = "OR" if index > 0 else ""
condition_sql += f"{joining} {this_entity_condition_sql}"
# wrap in smooths to constrain the scope of the OR
condition_sql = f"( {condition_sql} )"
params = {**params, **this_entity_filter_params}
params = {**params, "event_names": list(event_names_to_filter)}
if len(event_names_to_filter) == 0:
# using "All events"
having_conditions = ""
having_select = ""
else:
having_conditions = "AND hasAll(event_names, %(event_names)s)"
having_select = """
-- select the unique events in this session to support filtering sessions by presence of an event
groupUniqArray(event) as event_names,"""
return SummaryEventFiltersSQL(
having_conditions=having_conditions,
having_select=having_select,
where_conditions=f"AND {condition_sql}" if condition_sql else "",
params=params,
)
# We want to select events beyond the range of the recording to handle the case where
# a recording spans the time boundaries
@cached_property
def _get_events_timestamp_clause(self) -> Tuple[str, Dict[str, Any]]:
timestamp_clause = ""
timestamp_params = {}
if self._filter.date_from:
timestamp_clause += "\nAND timestamp >= %(event_start_time)s"
timestamp_params["event_start_time"] = self._filter.date_from - timedelta(
hours=12
)
if self._filter.date_to:
timestamp_clause += "\nAND timestamp <= %(event_end_time)s"
timestamp_params["event_end_time"] = self._filter.date_to + timedelta(
hours=12
)
return timestamp_clause, timestamp_params
def get_query(self, select_event_ids: bool = False) -> Tuple[str, Dict[str, Any]]:
if not self._determine_should_join_events():
return "", {}
base_params = {
"team_id": self._team_id,
"clamped_to_storage_ttl": (datetime.now() - timedelta(days=self.ttl_days)),
}
_, recording_start_time_params = _get_recording_start_time_clause(self._filter)
(
provided_session_ids_clause,
provided_session_ids_params,
) = _get_filter_by_provided_session_ids_clause(
recording_filters=self._filter, column_name="$session_id"
)
event_filters = self.build_event_filters
event_filters_params = event_filters.params
(
events_timestamp_clause,
events_timestamp_params,
) = self._get_events_timestamp_clause
# these will be applied to the events table,
# so we only want property filters that make sense in that context
prop_query, prop_params = self._get_prop_groups(
PropertyGroup(
type=PropertyOperatorType.AND,
values=[
g
for g in self._filter.property_groups.flat
if (g.type == "hogql" and "person.properties" not in g.key)
or (
g.type != "hogql"
and "cohort" not in g.type
and g.type != "person"
)
],
),
person_id_joined_alias=f"{self.DISTINCT_ID_TABLE_ALIAS}.person_id",
)
(
persons_join,
persons_select_params,
persons_sub_query,
) = self._persons_join_or_subquery(event_filters, prop_query)
return (
self._raw_events_query.format(
select_event_ids="groupArray(uuid) as event_ids,"
if select_event_ids
else "",
event_filter_where_conditions=event_filters.where_conditions,
event_filter_having_events_condition=event_filters.having_conditions,
event_filter_having_events_select=event_filters.having_select,
events_timestamp_clause=events_timestamp_clause,
prop_filter_clause=prop_query,
provided_session_ids_clause=provided_session_ids_clause,
persons_join=persons_join,
persons_sub_query=persons_sub_query,
),
{
**base_params,
**recording_start_time_params,
**provided_session_ids_params,
**events_timestamp_params,
**event_filters_params,
**prop_params,
**persons_select_params,
},
)
def _persons_join_or_subquery(self, event_filters, prop_query):
persons_select, persons_select_params = PersonsQuery(
filter=self._filter, team=self._team
).get_query()
persons_join = ""
persons_sub_query = ""
if persons_select:
# we want to join as infrequently as possible so only join if there are filters that expect it
if (
"person_props" in prop_query
or "pdi.person_id" in prop_query
or "person_props" in event_filters.where_conditions
):
persons_join = (
f"JOIN ({persons_select}) as pdi on pdi.distinct_id = e.distinct_id"
)
else:
persons_sub_query = f"AND e.distinct_id in (select distinct_id from ({persons_select}) as events_persons_sub_query)"
return persons_join, persons_select_params, persons_sub_query
@cached_property
def _get_person_id_clause(self) -> Tuple[str, Dict[str, Any]]:
person_id_clause = ""
person_id_params = {}
if self._filter.person_uuid:
person_id_clause = "AND person_id = %(person_uuid)s"
person_id_params = {"person_uuid": self._filter.person_uuid}
return person_id_clause, person_id_params
def matching_events(self) -> List[str]:
self._filter.hogql_context.person_on_events_mode = PersonOnEventsMode.DISABLED
query, query_params = self.get_query(select_event_ids=True)
query_results = sync_execute(
query, {**query_params, **self._filter.hogql_context.values}
)
results = [row[0] for row in query_results]
# flatten and return results
return [item for sublist in results for item in sublist]
class SessionRecordingListFromReplaySummary(EventQuery):
# we have to implement this from EventQuery but don't need it
def _determine_should_join_distinct_ids(self) -> None:
pass
_filter: SessionRecordingsFilter
SESSION_RECORDINGS_DEFAULT_LIMIT = 50
def __init__(
self,
**kwargs,
):
super().__init__(
**kwargs,
)
@property
def ttl_days(self):
return ttl_days(self._team)
_session_recordings_query: str = """
SELECT
s.session_id,
any(s.team_id),
any(s.distinct_id),
min(s.min_first_timestamp) as start_time,
max(s.max_last_timestamp) as end_time,
dateDiff('SECOND', start_time, end_time) as duration,
argMinMerge(s.first_url) as first_url,
sum(s.click_count),
sum(s.keypress_count),
sum(s.mouse_activity_count),
sum(s.active_milliseconds)/1000 as active_seconds,
duration-active_seconds as inactive_seconds,
sum(s.console_log_count) as console_log_count,
sum(s.console_warn_count) as console_warn_count,
sum(s.console_error_count) as console_error_count
FROM session_replay_events s
WHERE s.team_id = %(team_id)s
-- regardless of what other filters are applied
-- limit by storage TTL
AND s.min_first_timestamp >= %(clamped_to_storage_ttl)s
-- we can filter on the pre-aggregated timestamp columns
-- because any not-the-lowest min value is _more_ greater than the min value
-- and any not-the-highest max value is _less_ lower than the max value
AND s.min_first_timestamp >= %(start_time)s
AND s.max_last_timestamp <= %(end_time)s
{persons_sub_query}
{events_sub_query}
{provided_session_ids_clause}
GROUP BY session_id
HAVING 1=1 {duration_clause} {console_log_clause}
ORDER BY start_time DESC
LIMIT %(limit)s OFFSET %(offset)s
"""
@staticmethod
def _data_to_return(results: List[Any]) -> List[Dict[str, Any]]:
default_columns = [
"session_id",
"team_id",
"distinct_id",
"start_time",
"end_time",
"duration",
"first_url",
"click_count",
"keypress_count",
"mouse_activity_count",
"active_seconds",
"inactive_seconds",
"console_log_count",
"console_warn_count",
"console_error_count",
]
return [
{
**dict(zip(default_columns, row[: len(default_columns)])),
}
for row in results
]
def _paginate_results(self, session_recordings) -> SessionRecordingQueryResult:
more_recordings_available = False
if len(session_recordings) > self.limit:
more_recordings_available = True
session_recordings = session_recordings[0 : self.limit]
return SessionRecordingQueryResult(
session_recordings, more_recordings_available
)
def run(self) -> SessionRecordingQueryResult:
self._filter.hogql_context.person_on_events_mode = PersonOnEventsMode.DISABLED
query, query_params = self.get_query()
query_results = sync_execute(
query, {**query_params, **self._filter.hogql_context.values}
)
session_recordings = self._data_to_return(query_results)
return self._paginate_results(session_recordings)
@property
def limit(self):
return self._filter.limit or self.SESSION_RECORDINGS_DEFAULT_LIMIT
def get_query(self) -> Tuple[str, Dict[str, Any]]:
offset = self._filter.offset or 0
base_params = {
"team_id": self._team_id,
"limit": self.limit + 1,
"offset": offset,
"clamped_to_storage_ttl": (datetime.now() - timedelta(days=self.ttl_days)),
}
_, recording_start_time_params = _get_recording_start_time_clause(self._filter)
(
provided_session_ids_clause,
provided_session_ids_params,
) = _get_filter_by_provided_session_ids_clause(recording_filters=self._filter)
duration_clause, duration_params = self.duration_clause(
self._filter.duration_type_filter
)
console_log_clause = self._get_console_log_clause(
self._filter.console_logs_filter
)
events_select, events_join_params = SessionIdEventsQuery(
team=self._team,
filter=self._filter,
).get_query()
if events_select:
events_select = f"AND s.session_id in (select `$session_id` as session_id from ({events_select}) as session_events_sub_query)"
persons_select, persons_select_params = PersonsQuery(
filter=self._filter, team=self._team
).get_query()
if persons_select:
persons_select = f"AND s.distinct_id in (select distinct_id from ({persons_select}) as session_persons_sub_query)"
return (
self._session_recordings_query.format(
duration_clause=duration_clause,
provided_session_ids_clause=provided_session_ids_clause,
console_log_clause=console_log_clause,
persons_sub_query=persons_select,
events_sub_query=events_select,
),
{
**base_params,
**events_join_params,
**recording_start_time_params,
**duration_params,
**provided_session_ids_params,
**persons_select_params,
},
)
def duration_clause(
self,
duration_filter_type: Literal["duration", "active_seconds", "inactive_seconds"],
) -> Tuple[str, Dict[str, Any]]:
duration_clause = ""
duration_params = {}
if self._filter.recording_duration_filter:
if self._filter.recording_duration_filter.operator == "gt":
operator = ">"
else:
operator = "<"
duration_clause = (
"\nAND {duration_type} {operator} %(recording_duration)s".format(
duration_type=duration_filter_type, operator=operator
)
)
duration_params = {
"recording_duration": self._filter.recording_duration_filter.value,
}
return duration_clause, duration_params
@staticmethod
def _get_console_log_clause(
console_logs_filter: List[Literal["error", "warn", "log"]]
) -> str:
filters = [f"console_{log}_count > 0" for log in console_logs_filter]
return f"AND ({' OR '.join(filters)})" if filters else ""
|
Code | OpeningGuide | import atexit
import os
import sqlite3
import Code.SQL.DBF as SQLDBF
import LCEngine4 as LCEngine
from Code import AperturasStd, Books, ControlPosicion, PGNreader, Util, VarGen
from Code.QT import QTUtil2, QTVarios
class UnMove:
def __init__(self, bookGuide, father):
self.bookGuide = bookGuide
self.dbAnalisis = bookGuide.dbAnalisis
self._rowid = None
self._father = father
self._pv = ""
self._xpv = ""
self._nag = 0
self._adv = 0
self._comment = ""
self._pos = 0
self._xdata = {}
self._graphics = ""
self._mark = ""
self._children = []
self._item = None
if father:
self._siBlancas = not father.siBlancas()
self._numJugada = father.numJugada() + (1 if self._siBlancas else 0)
else: # root
self._siBlancas = False
self._numJugada = 0
self._fen = ""
self._pgn = "" # set al crear el pv
self.readedEXT = False
def __str__(self):
return "%s %s %s %s %s %s %s" % (
self.rowid(),
self.father().rowid(),
self.pv(),
self.xpv(),
self.siBlancas(),
self.numJugada(),
self.fen(),
)
def rowid(self, valor=None):
if valor is not None:
self._rowid = valor
return self._rowid
def siBlancas(self):
return self._siBlancas
def numJugada(self):
return self._numJugada
def father(self):
return self._father
def pv(self, valor=None):
if valor is not None:
self._pv = valor
return self._pv
def xpv(self, valor=None):
if valor is not None:
self._xpv = valor
return self._xpv
def mark(self, valor=None, siLeyendo=False):
if valor is not None:
ant = self._mark
self._mark = valor
if not siLeyendo:
if ant != valor:
self.bookGuide.pteGrabar(self)
self.bookGuide.actualizaBookmark(self, len(valor) > 0)
return self._mark
def graphics(self, valor=None, siLeyendo=False):
if valor is None:
if not self.readedEXT:
self.readEXT()
else:
ant = self._graphics
self._graphics = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
return self._graphics
def nag(self, valor=None, siLeyendo=False):
if valor is not None:
ant = self._nag
self._nag = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
return self._nag
def adv(self, valor=None, siLeyendo=False):
if valor is not None:
ant = self._adv
self._adv = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
return self._adv
def comment(self, valor=None, siLeyendo=False):
if valor is None:
self.readEXT()
else:
ant = self._comment
self._comment = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
return self._comment
def commentLine(self):
c = self.comment()
if c:
li = c.split("\n")
c = li[0]
if len(li) > 1:
c += "..."
return c
def xdata(self, valor=None, siLeyendo=False):
if valor is None:
self.readEXT()
else:
ant = self._xdata
self._xdata = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
self._xdata = valor
return self._xdata
def pos(self, valor=None, siLeyendo=False):
if valor is not None:
ant = self._pos
self._pos = valor
if not siLeyendo and ant != valor:
self.bookGuide.pteGrabar(self)
return self._pos
def fen(self, valor=None):
if valor is not None:
self._fen = valor
self.bookGuide.setTransposition(self)
return self._fen
def fenM2(self):
fen = self._fen
sp2 = fen.rfind(" ", 0, fen.rfind(" "))
return fen[:sp2]
def transpositions(self):
return self.bookGuide.getTranspositions(self)
def fenBase(self):
return self._father._fen
def pgn(self):
if not self._pgn:
pv = self._pv
d, h, c = pv[:2], pv[2:4], pv[4:]
cp = ControlPosicion.ControlPosicion()
cp.leeFen(self._father.fen())
self._pgn = cp.pgnSP(d, h, c)
cp.mover(d, h, c)
self._fen = cp.fen()
return self._pgn
def pgnEN(self):
cp = ControlPosicion.ControlPosicion()
cp.leeFen(self._father.fen())
pv = self._pv
d, h, c = pv[:2], pv[2:4], pv[4:]
return cp.pgn(d, h, c)
def pgnNum(self):
if self._siBlancas:
return "%d.%s" % (self._numJugada, self.pgn())
return self.pgn()
def item(self, valor=None):
if valor is not None:
self._item = valor
return self._item
def children(self):
return self._children
def addChildren(self, move):
self._children.append(move)
self._children = sorted(self._children, key=lambda uno: uno._pos)
def delChildren(self, move):
for n, mv in enumerate(self._children):
if move == mv:
del self._children[n]
return
def brothers(self):
li = []
for mv in self._father.children():
if mv.pv() != self.pv():
li.append(mv)
return li
def analisis(self):
return self.dbAnalisis.move(self)
def etiPuntos(self):
rm = self.dbAnalisis.move(self)
if rm:
return rm.abrTextoBase()
else:
return ""
def historia(self):
li = []
p = self
while True:
li.insert(0, p)
if not p.father():
break
p = p.father()
return li
def allPV(self):
return LCEngine.xpv2pv(self._xpv)
def allPGN(self):
li = []
for mv in self.historia():
if mv._pv:
if mv.siBlancas():
li.append("%d." % mv.numJugada())
li.append(mv.pgn())
return " ".join(li)
def readEXT(self):
if self.readedEXT:
return
self.readedEXT = True
self.bookGuide.readEXT(self)
class OpeningGuide:
def __init__(self, wowner, nomFichero=None):
self.configuracion = VarGen.configuracion
siGenerarStandard = False
if nomFichero is None:
nomFichero = self.configuracion.ficheroBookGuide
if not os.path.isfile(nomFichero):
siGenerarStandard = "Standard opening guide" in nomFichero
self.name = os.path.basename(nomFichero)[:-4]
self.ultPos = 0
self.dicPtes = {}
self.nomFichero = nomFichero
self.conexion = sqlite3.connect(nomFichero)
self.conexion.text_factory = lambda x: unicode(x, "utf-8", "ignore")
atexit.register(self.cerrar)
self.tablaDatos = "GUIDE"
self.checkInitBook(wowner, siGenerarStandard)
self.transpositions = {}
self.bookmarks = []
self.dbAnalisis = DBanalisis()
def pathGuide(self, nameGuide):
return Util.dirRelativo(
os.path.join(self.configuracion.carpeta, nameGuide + ".pgo")
)
def getOtras(self):
li = Util.listdir(self.configuracion.carpeta)
lwbase = self.name.lower()
liresp = []
for uno in li:
lw = uno.name.lower()
if lw.endswith(".pgo"):
if lwbase != lw[:-4]:
liresp.append(uno.name[:-4])
return liresp
def getTodas(self):
li = self.getOtras()
li.append(self.name)
return li
def changeTo(self, wowner, nomGuide):
self.grabar()
nomFichero = self.pathGuide(nomGuide)
self.name = nomGuide
self.ultPos = 0
self.dicPtes = {}
self.nomFichero = nomFichero
self.conexion.close()
self.conexion = sqlite3.connect(nomFichero)
self.conexion.text_factory = lambda x: unicode(x, "utf-8", "ignore")
atexit.register(self.cerrar)
self.checkInitBook(wowner, False)
self.transpositions = {}
self.bookmarks = []
self.root = UnMove(self, None)
self.root._fen = ControlPosicion.FEN_INICIAL
self.readAllDB()
self.configuracion.ficheroBookGuide = nomFichero
self.configuracion.graba()
def copyTo(self, otraGuide):
self.grabar()
otroFichero = self.pathGuide(otraGuide)
Util.copiaFichero(self.nomFichero, otroFichero)
def renameTo(self, wowner, otraGuide):
self.grabar()
self.conexion.close()
otroFichero = self.pathGuide(otraGuide)
Util.renombraFichero(self.nomFichero, otroFichero)
self.changeTo(wowner, otraGuide)
def removeOther(self, otraGuide):
self.grabar()
otroFichero = self.pathGuide(otraGuide)
Util.borraFichero(otroFichero)
def appendFrom(self, wowner, otraGuide):
self.grabar()
otroFichero = self.pathGuide(otraGuide)
otraConexion = sqlite3.connect(otroFichero)
otraConexion.text_factory = lambda x: unicode(x, "utf-8", "ignore")
cursor = otraConexion.cursor()
cursor.execute("pragma table_info(%s)" % self.tablaDatos)
liCamposOtra = cursor.fetchall()
cursor.close()
if not liCamposOtra:
return False
st = set()
for x in liCamposOtra:
st.add(x[1]) # nombre
liselect = (
"XPV",
"PV",
"NAG",
"ADV",
"COMMENT",
"FEN",
"MARK",
"GRAPHICS",
"XDATA",
)
libasic = ("XPV", "PV", "FEN")
li = []
for x in liselect:
if x not in st:
if x in libasic:
otraConexion.close()
QTUtil2.mensError(wowner, _("This guide file is not valid"))
return False
else:
li.append(x)
select = ",".join(li)
dbfOtra = SQLDBF.DBF(otraConexion, self.tablaDatos, select)
dbfOtra.leer()
reccount = dbfOtra.reccount()
bp = QTUtil2.BarraProgreso(wowner, otraGuide, "", reccount).mostrar()
liReg = []
for recno in range(reccount):
bp.pon(recno)
if bp.siCancelado():
break
dbfOtra.goto(recno)
reg = dbfOtra.registroActual()
liReg.append(reg)
dbfOtra.cerrar()
otraConexion.close()
def dispatch(recno):
bp.pon(recno)
if liReg:
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, select)
dbf.insertarLista(liReg, dispatch)
dbf.cerrar()
bp.cerrar()
return len(liReg) > 0
def generarStandard(self, ventana, siBasic):
oLista = AperturasStd.apTrain
dic = oLista.dic
titulo = _("Openings")
tmpBP2 = QTUtil2.BarraProgreso2(ventana, titulo)
tf = len(dic)
tmpBP2.ponTotal(1, tf)
tmpBP2.ponRotulo(1, "1. " + _X(_("Reading %1"), titulo))
tmpBP2.ponTotal(2, tf)
tmpBP2.ponRotulo(2, "")
tmpBP2.mostrar()
liRegs = []
dRegs = {} # se guarda una lista con los pv, para determinar el padre
for nR, k in enumerate(oLista.dic):
tmpBP2.pon(1, nR)
tmpBP2.siCancelado()
ae = oLista.dic[k]
if siBasic and not ae.siBasic:
continue
liPV = ae.a1h8.split(" ")
ult = len(liPV) - 1
seqFather = ""
cp = ControlPosicion.ControlPosicion()
cp.posInicial()
for pos, pv in enumerate(liPV):
desde, hasta, coronacion = pv[:2], pv[2:4], pv[4:]
seq = seqFather + LCEngine.pv2xpv(pv)
cp.mover(desde, hasta, coronacion)
if seq not in dRegs:
reg = SQLDBF.Almacen()
reg.XPV = seq
reg.PV = pv
reg.FEN = cp.fen()
reg.COMMENT = ae.trNombre if pos == ult else ""
self.ultPos += 1
reg.POS = self.ultPos
liRegs.append(reg)
dRegs[seq] = reg
seqFather = seq
tmpBP2.ponRotulo(2, "2. " + _X(_("Converting %1"), titulo))
tmpBP2.ponTotal(2, len(liRegs))
select = "XPV,PV,COMMENT,FEN,POS"
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, select)
def dispatch(num):
tmpBP2.pon(2, num)
tmpBP2.siCancelado()
dbf.insertarLista(liRegs, dispatch)
dbf.cerrar()
tmpBP2.cerrar()
def creaTabla(self):
cursor = self.conexion.cursor()
sql = (
"CREATE TABLE %s( XPV TEXT UNIQUE,PV VARCHAR(5),NAG INTEGER,ADV INTEGER,COMMENT TEXT,"
"FEN VARCHAR,MARK VARCHAR, POS INTEGER,GRAPHICS TEXT,XDATA BLOB);"
) % self.tablaDatos
cursor.execute(sql)
self.conexion.commit()
cursor.close()
def checkInitBook(self, wowner, siGenerarStandard):
cursor = self.conexion.cursor()
cursor.execute("pragma table_info(%s)" % self.tablaDatos)
liCampos = cursor.fetchall()
cursor.close()
if not liCampos:
self.creaTabla()
if siGenerarStandard:
self.generarStandard(wowner, True)
def grabarPGN(self, ventana, ficheroPGN, maxDepth):
select = "XPV,PV,COMMENT,NAG,ADV,FEN,POS"
SQLDBF.DBF(self.conexion, self.tablaDatos, select)
erroneos = duplicados = importados = 0
dlTmp = QTVarios.ImportarFicheroPGN(ventana)
dlTmp.hideDuplicados()
dlTmp.show()
select = "XPV,PV,COMMENT,NAG,ADV,FEN,POS"
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, select)
dnag = {"!!": 3, "!": 1, "?": 2, "??": 4, "!?": 5, "?!": 6}
n = 0
liReg = []
for n, g in enumerate(PGNreader.readGames(ficheroPGN), 1):
if not dlTmp.actualiza(n, erroneos, duplicados, importados):
break
if g.erroneo:
erroneos += 1
continue
if not g.moves:
erroneos += 1
continue
liReg = []
def addMoves(moves, depth, seq):
for mv in moves.liMoves:
if depth > maxDepth:
break
seqM1 = seq
pv = mv.pv
seq += LCEngine.pv2xpv(pv)
reg = SQLDBF.Almacen()
reg.PV = pv
reg.XPV = seq
reg.COMMENT = "\n".join(mv.comentarios)
reg.FEN = mv.fen
reg.NAG = 0
reg.ADV = 0
self.ultPos += 1
reg.POS = self.ultPos
for critica in mv.criticas:
if critica.isdigit():
t = int(critica)
if t in (4, 2, 1, 3, 5, 6):
reg.NAG = t
elif t in (11, 14, 15, 16, 17, 18, 19):
reg.ADV = t
else:
if critica in dnag:
reg.NAG = dnag[critica]
liReg.append(reg)
if mv.variantes:
for variante in mv.variantes:
addMoves(variante, depth, seqM1)
depth += 1
addMoves(g.moves, 1, "")
if liReg:
dbf.insertarLista(liReg, None)
dbf.cerrar()
dlTmp.actualiza(n, erroneos, duplicados, importados)
dlTmp.ponContinuar()
return len(liReg) > 0
def grabarPolyglot(self, ventana, ficheroBIN, depth, whiteBest, blackBest):
titulo = _("Import a polyglot book")
tmpBP2 = QTUtil2.BarraProgreso2(ventana, titulo)
tmpBP2.ponTotal(1, 1)
tmpBP2.ponRotulo(1, "1. " + _X(_("Reading %1"), os.path.basename(ficheroBIN)))
tmpBP2.ponTotal(2, 1)
tmpBP2.ponRotulo(2, "")
tmpBP2.mostrar()
basePos = self.ultPos
book = Books.Libro("P", ficheroBIN, ficheroBIN, True)
book.polyglot()
cp = ControlPosicion.ControlPosicion()
lireg = []
stFenM2 = set() # para que no se produzca un circulo vicioso
def hazFEN(fen, ply, seq):
plyN = ply + 1
siWhite = " w " in fen
siMax = False
if whiteBest:
siMax = siWhite
if blackBest:
siMax = siMax or not siWhite
liPV = book.miraListaPV(fen, siMax)
for pv in liPV:
cp.leeFen(fen)
cp.mover(pv[:2], pv[2:4], pv[4:])
fenN = cp.fen()
reg = SQLDBF.Almacen()
lireg.append(reg)
reg.PV = pv
seqN = seq + LCEngine.pv2xpv(pv)
reg.XPV = seqN
reg.COMMENT = ""
reg.NAG = 0
reg.FEN = fenN
reg.ADV = 0
self.ultPos += 1
reg.POS = self.ultPos
tmpBP2.ponTotal(1, self.ultPos - basePos)
tmpBP2.pon(1, self.ultPos - basePos)
if plyN < depth:
fenM2 = cp.fenM2()
if fenM2 not in stFenM2:
stFenM2.add(fenM2)
hazFEN(fenN, plyN, seqN)
hazFEN(ControlPosicion.FEN_INICIAL, 0, "")
select = "XPV,PV,COMMENT,NAG,ADV,FEN,POS"
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, select)
tmpBP2.ponTotal(2, len(lireg))
tmpBP2.ponRotulo(2, _("Writing..."))
def dispatch(num):
tmpBP2.pon(2, num)
dbf.insertarLista(lireg, dispatch)
dbf.cerrar()
tmpBP2.cerrar()
return len(lireg) > 0
def reset(self):
self.grabar()
self.dicPtes = {}
self.root = UnMove(self, None)
self.root._fen = ControlPosicion.FEN_INICIAL
self.readAllDB()
def readAllDB(self):
self.transpositions = {}
self.bookmarks = []
self.ultPos = 0
select = "ROWID,XPV,PV,NAG,ADV,FEN,MARK,POS"
orden = "XPV"
condicion = ""
dbf = SQLDBF.DBFT(self.conexion, self.tablaDatos, select, condicion, orden)
dbf.leer()
dicMoves = {}
dicMoves[""] = self.root
for recno in range(dbf.reccount()):
dbf.goto(recno)
xpv = dbf.XPV
pv = dbf.PV
if not pv:
self.root.rowid(dbf.ROWID)
continue
xpvfather = xpv[: -2 if len(pv) == 4 else -3]
if xpvfather in dicMoves:
father = dicMoves[xpvfather]
mv = UnMove(self, father)
mv.pv(pv)
mv.xpv(xpv)
mv.fen(dbf.FEN)
mv.rowid(dbf.ROWID)
mv.nag(dbf.NAG, True)
mv.adv(dbf.ADV, True)
mark = dbf.MARK
if mark:
self.bookmarks.append(mv)
mv.mark(dbf.MARK, True)
mv.pos(dbf.POS, True)
if dbf.POS >= self.ultPos:
self.ultPos = dbf.POS
dicMoves[xpv] = mv
father.addChildren(mv)
dbf.cerrar()
def setTransposition(self, move):
fenM2 = move.fenM2()
if fenM2 not in self.transpositions:
self.transpositions[fenM2] = [move]
else:
li = self.transpositions[fenM2]
if move not in li:
li.append(move)
def getTranspositions(self, move):
li = self.transpositions.get(move.fenM2(), [])
if len(li) <= 1:
return []
n = li.index(move)
li = li[:]
del li[n]
return li
def getMovesFenM2(self, fenM2):
return self.transpositions.get(fenM2, None)
def actualizaBookmark(self, move, siPoner):
siEsta = move in self.bookmarks
if siEsta:
if not siPoner:
del self.bookmarks[self.bookmarks.index(move)]
else:
if siPoner:
self.bookmarks.append(move)
def readEXT(self, move):
select = "COMMENT,GRAPHICS,XDATA"
condicion = "XPV='%s'" % move.xpv()
dbf = SQLDBF.DBFT(self.conexion, self.tablaDatos, select, condicion)
dbf.leer()
if dbf.reccount():
dbf.goto(0)
move.comment(dbf.COMMENT, True)
move.graphics(dbf.GRAPHICS, True)
move.xdata(Util.blob2var(dbf.XDATA), True)
dbf.cerrar()
def pteGrabar(self, move):
huella = move.xpv()
if huella not in self.dicPtes:
self.dicPtes[huella] = move
if len(self.dicPtes) > 5:
self.grabar()
def mixTable(self, tableFrom, tableTo):
self.grabar()
nameFrom = "DATA%d" % tableFrom
nameTo = "DATA%d" % tableTo
cursor = self.conexion.cursor()
cursor.execute("SELECT ROWID,XPV FROM %s" % nameFrom)
liValores = cursor.fetchall()
for rowid, xpv in liValores:
cursor.execute('SELECT ROWID FROM %s WHERE XPV="%s"' % (nameTo, xpv))
li = cursor.fetchone()
if li:
rowidTo = li[0]
sql = "DELETE FROM %s WHERE rowid = %d" % (nameTo, rowidTo)
cursor.execute(sql)
sql = "INSERT INTO %s SELECT * FROM %s WHERE %s.ROWID = %d;" % (
nameTo,
nameFrom,
nameFrom,
rowid,
)
cursor.execute(sql)
self.conexion.commit()
cursor.close()
def grabarFichSTAT(self, nomGuide, fich):
# Para convertir datos de games a bookGuide
self.changeTo(None, nomGuide)
f = open(fich, "rb")
liRegs = []
for linea in f:
linea = linea.strip()
xpv, pv, fen = linea.split("|")
reg = SQLDBF.Almacen()
reg.XPV = xpv
reg.PV = pv
reg.FEN = fen
self.ultPos += 1
reg.POS = self.ultPos
liRegs.append(reg)
select = "XPV,PV,FEN,POS"
dbf = SQLDBF.DBFT(self.conexion, self.tablaDatos, select)
def dispatch(num):
pass
dbf.insertarLista(liRegs, dispatch)
dbf.cerrar()
def cerrar(self):
if self.conexion:
self.conexion.close()
self.conexion = None
self.dbAnalisis.cerrar()
def grabar(self):
if len(self.dicPtes) == 0:
return
dic = self.dicPtes
self.dicPtes = {}
# Creamos una tabla de trabajo
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, "")
for k, uno in dic.items():
reg = SQLDBF.Almacen()
reg.XPV = uno.xpv()
reg.PV = uno.pv()
reg.NAG = uno.nag()
reg.ADV = uno.adv()
reg.COMMENT = uno.comment()
reg.POS = uno.pos()
reg.FEN = uno.fen()
reg.MARK = uno.mark()
reg.GRAPHICS = uno.graphics()
reg.XDATA = Util.var2blob(uno.xdata())
if uno.rowid() is None:
xid = dbf.insertarSoloReg(reg)
uno.rowid(xid)
else:
dbf.modificarROWID(uno.rowid(), reg)
dbf.cerrar()
def dameMovimiento(self, father, pv):
mv = UnMove(self, father)
xpv = father.xpv() + LCEngine.pv2xpv(pv)
mv.xpv(xpv)
mv.pv(pv)
cp = ControlPosicion.ControlPosicion()
cp.leeFen(father.fen())
cp.moverPV(pv)
mv.fen(cp.fen())
self.ultPos += 1
mv.pos(self.ultPos)
father.addChildren(mv)
self.pteGrabar(mv)
return mv
def borrar(self, uno):
liBorrados = [uno]
def allChildren(li, uno):
for x in uno.children():
li.append(x.rowid())
liBorrados.append(x)
allChildren(li, x)
liRowid = []
if uno.rowid():
liRowid.append(uno.rowid())
allChildren(liRowid, uno)
if liRowid:
dbf = SQLDBF.DBF(self.conexion, self.tablaDatos, "")
dbf.borrarListaRaw(liRowid)
if len(liRowid) > 10:
dbf.pack()
dbf.cerrar()
liQuitarTrasposition = []
for mov in liBorrados:
if mov in self.bookmarks:
del self.bookmarks[self.bookmarks.index(mov)]
fenM2 = mov.fenM2()
li = self.transpositions[fenM2]
if len(li) <= 1:
del self.transpositions[fenM2]
else:
del li[li.index(mov)]
if len(li) == 1:
xm = li[0]
if xm not in liBorrados:
liQuitarTrasposition.append(xm)
return liBorrados, liQuitarTrasposition
def allLines(self):
rt = self.root
liT = []
def uno(mv):
li = mv.children()
if li:
for mv in li:
uno(mv)
else:
liT.append(mv.historia())
uno(rt)
return liT
class DBanalisis:
def __init__(self):
self.db = Util.DicSQL(
VarGen.configuracion.ficheroAnalisisBookGuide,
tabla="analisis",
maxCache=1024,
)
def cerrar(self):
self.db.close()
def lista(self, fenM2):
dic = self.db[fenM2]
if dic:
lista = dic.get("LISTA", None)
activo = dic.get("ACTIVO", None)
else:
lista = None
activo = None
return lista, activo
def mrm(self, fenM2):
dic = self.db[fenM2]
if dic:
lista = dic.get("LISTA", None)
if lista:
nactive = dic.get("ACTIVO", None)
if nactive is not None:
return lista[nactive]
return None
def move(self, move):
fenM2 = move.father().fenM2()
dic = self.db[fenM2]
if dic:
numactivo = dic.get("ACTIVO", None)
if numactivo is not None:
lista = dic.get("LISTA", None)
if lista:
if 0 < numactivo >= len(lista):
numactivo = 0
dic["ACTIVO"] = 0
self.db[fenM2] = dic
analisis = lista[numactivo]
if analisis:
rm, k = analisis.buscaRM(move.pv())
return rm
return None
def getAnalisis(self, fenM2):
dic = self.db[fenM2]
if not dic:
dic = {"ACTIVO": None, "LISTA": []}
return dic
def nuevo(self, fenM2, analisis):
dic = self.getAnalisis(fenM2)
li = dic["LISTA"]
li.append(analisis)
dic["ACTIVO"] = len(li) - 1
self.db[fenM2] = dic
def pon(self, fenM2, numActivo):
dic = self.getAnalisis(fenM2)
dic["ACTIVO"] = numActivo
self.db[fenM2] = dic
def activo(self, fenM2):
dic = self.getAnalisis(fenM2)
return dic["ACTIVO"]
def quita(self, fenM2, num):
dic = self.getAnalisis(fenM2)
li = dic["LISTA"]
del li[num]
numActivo = dic["ACTIVO"]
if numActivo is not None:
if numActivo == num:
numActivo = None
elif numActivo > num:
numActivo -= 1
dic["ACTIVO"] = numActivo
self.db[fenM2] = dic
|
pdu | qa_time_delta | #!/usr/bin/env python
#
# Copyright 2021 NTESS LLC.
#
# This file is part of GNU Radio
#
# SPDX-License-Identifier: GPL-3.0-or-later
#
#
import time
import numpy as np
import pmt
from gnuradio import blocks, gr, gr_unittest, pdu
class qa_time_delta(gr_unittest.TestCase):
def setUp(self):
self.tb = gr.top_block()
self.time_delta = pdu.time_delta(
pmt.intern("sys time delta (ms)"), pmt.intern("system_time")
)
self.debug = blocks.message_debug()
self.tb.msg_connect((self.time_delta, "pdu"), (self.debug, "store"))
def tearDown(self):
self.tb = None
def test_001_invalid_a(self):
self.tb.start()
self.time_delta.to_basic_block()._post(
pmt.intern("pdu"), pmt.intern("NOT A PDU")
)
time.sleep(0.01) # short delay to ensure the message is processed
self.tb.stop()
self.tb.wait()
# nothing should be produced in this case
self.assertEqual(0, self.debug.num_messages())
def test_001_invalid_b(self):
in_data = [
0,
0,
0,
0,
0,
0,
0,
0,
1,
1,
1,
1,
1,
1,
1,
1,
0,
1,
0,
1,
0,
1,
0,
1,
]
meta = pmt.dict_add(pmt.make_dict(), pmt.intern("sam"), pmt.from_double(25.1))
in_pdu = pmt.cons(meta, pmt.init_c32vector(len(in_data), in_data))
# set up fg
self.tb.start()
self.time_delta.to_basic_block()._post(pmt.intern("pdu"), in_pdu)
time.sleep(0.01) # short delay to ensure the message is processed
self.tb.stop()
self.tb.wait()
# nothing should be produced in this case
self.assertEqual(0, self.debug.num_messages())
def test_002_normal(self):
tnow = time.time()
in_data = [
0,
0,
0,
0,
0,
0,
0,
0,
1,
1,
1,
1,
1,
1,
1,
1,
0,
1,
0,
1,
0,
1,
0,
1,
]
meta = pmt.dict_add(
pmt.make_dict(), pmt.intern("system_time"), pmt.from_double(tnow - 10.0)
)
in_pdu = pmt.cons(meta, pmt.init_c32vector(len(in_data), in_data))
# set up fg
self.tb.start()
self.time_delta.to_basic_block()._post(pmt.intern("pdu"), in_pdu)
self.waitFor(
lambda: self.debug.num_messages() == 1, timeout=1.0, poll_interval=0.01
)
self.tb.stop()
self.tb.wait()
# check data
self.assertEqual(1, self.debug.num_messages())
a_meta = pmt.car(self.debug.get_message(0))
time_tag = pmt.dict_ref(a_meta, pmt.intern("system_time"), pmt.PMT_NIL)
delta_tag = pmt.dict_ref(a_meta, pmt.intern("sys time delta (ms)"), pmt.PMT_NIL)
self.assertAlmostEqual(tnow - 10.0, pmt.to_double(time_tag), delta=1e-6)
self.assertAlmostEqual(10000, pmt.to_double(delta_tag), delta=500)
if __name__ == "__main__":
gr_unittest.run(qa_time_delta)
|
decrypters | MultiloadCz | # -*- coding: utf-8 -*-
import re
from ..base.decrypter import BaseDecrypter
class MultiloadCz(BaseDecrypter):
__name__ = "MultiloadCz"
__type__ = "decrypter"
__version__ = "0.46"
__status__ = "testing"
__pattern__ = r"http://(?:[^/]*\.)?multiload\.cz/(stahnout|slozka)/.+"
__config__ = [
("enabled", "bool", "Activated", True),
("use_premium", "bool", "Use premium account if available", True),
("use_subfolder", "bool", "Save package to subfolder", True),
("subfolder_per_package", "bool", "Create a subfolder for each package", True),
("usedHoster", "str", "Prefered hoster list (bar-separated)", ""),
("ignoredHoster", "str", "Ignored hoster list (bar-separated)", ""),
]
__description__ = """Multiload.cz decrypter plugin"""
__license__ = "GPLv3"
__authors__ = [("zoidberg", "zoidberg@mujmail.cz")]
FOLDER_PATTERN = (
r'<form action="" method="get"><textarea.*?>([^>]*)</textarea></form>'
)
LINK_PATTERN = r'<p class="manager-server"><strong>(.+?)</strong></p><p class="manager-linky"><a href="(.+?)">'
def decrypt(self, pyfile):
self.data = self.load(pyfile.url)
if re.match(self.__pattern__, pyfile.url).group(1) == "slozka":
m = re.search(self.FOLDER_PATTERN, self.data)
if m is not None:
self.links.extend(m.group(1).split())
else:
m = re.findall(self.LINK_PATTERN, self.data)
if m is not None:
prefered_set = set(self.config.get("usedHoster").split("|"))
self.links.extend(x[1] for x in m if x[0] in prefered_set)
if not self.links:
ignored_set = set(self.config.get("ignoredHoster").split("|"))
self.links.extend(x[1] for x in m if x[0] not in ignored_set)
|
documents | models | """Entity objects for the Document Management applications.
TODO: move to an independent service / app.
"""
from __future__ import annotations
import itertools
import logging
import mimetypes
import threading
import uuid
from pathlib import Path
from typing import TYPE_CHECKING, Any, Collection, Iterator
import pkg_resources
import sqlalchemy as sa
import whoosh.fields as wf
from abilian.core.entities import Entity, db
from abilian.core.models import NOT_AUDITABLE, SEARCHABLE
from abilian.core.models.blob import Blob
from abilian.core.models.subjects import Group, User
from abilian.core.util import md5
from abilian.services.conversion import converter
from abilian.services.indexing import indexable_role
from abilian.services.security import Admin, Anonymous, InheritSecurity, security
from flask import current_app, json, url_for
from flask_login import current_user
from sqlalchemy.event import listen, listens_for
from sqlalchemy.ext.hybrid import hybrid_property
from sqlalchemy.orm import backref, foreign, relationship, remote
from sqlalchemy.orm.attributes import Event
from sqlalchemy.orm.session import Session
from sqlalchemy.schema import Column, ForeignKey, UniqueConstraint
from sqlalchemy.types import Integer, Text, UnicodeText
from sqlalchemy.util.langhelpers import symbol
from toolz import first
from whoosh.analysis import CharsetFilter, LowercaseFilter, RegexTokenizer
from whoosh.support.charset import accent_map
from . import tasks
from .lock import Lock
if TYPE_CHECKING:
from abilian.sbe.apps.communities.models import Community
logger = logging.getLogger(__package__)
__all__ = (
"CmisObject",
"Folder",
"Document",
"BaseContent",
"PathAndSecurityIndexable",
"icon_for",
"icon_url",
"setup_listener",
)
#: A Whoosh analyzer that folds accents and case.
accent_folder = RegexTokenizer() | LowercaseFilter() | CharsetFilter(accent_map)
ICONS_FOLDER = pkg_resources.resource_filename("abilian.sbe", "static/fileicons")
def icon_url(filename: str) -> str:
return url_for("abilian_sbe_static", filename=f"fileicons/{filename}")
def icon_exists(filename: str) -> bool:
return Path(ICONS_FOLDER, filename).is_file()
#
# Domain classes
#
class CmisObject(InheritSecurity, Entity):
"""(Abstract) Base class for CMIS objects."""
# normally set by communities.models.community_content,
# but we have to fix a circ.dep to use it
is_community_content = True
__tablename__ = "cmisobject"
__indexable__ = False
__index_to__ = (
("community.id", ("community_id",)),
#
("community.slug", ("community_slug",)),
)
_title = Column("title", UnicodeText, nullable=False, default="")
description = Column(
UnicodeText,
nullable=False,
default="",
info=SEARCHABLE | {"index_to": ("description", "text")},
)
_parent_id = Column(Integer, ForeignKey("cmisobject.id"), nullable=True)
# no duplicate name in same folder
__table_args__ = (UniqueConstraint("_parent_id", "title"),)
# Set in concrete classes
sbe_type: str = None
# Convenience default values
content_length = 0
name: str
def __init__(self, *args, **kwargs):
# ensure 'title' prevails over 'name'
if "title" in kwargs and "name" in kwargs:
title = kwargs.get("title")
name = kwargs.get("name")
if title is None:
del kwargs["title"]
elif name != title:
kwargs["name"] = title
Entity.__init__(self, *args, **kwargs)
# title is defined has an hybrid property to allow 2 way sync name <->
# title
@hybrid_property
def title(self: str) -> str:
return self._title
@title.setter
def title(self, title: str):
# set title before setting name, so that we don't enter an infinite loop
# with _cmis_sync_name_title
self._title = title
if self.name != title:
self.name = title
def clone(
self, title: str | None = None, parent: Folder | None = None
) -> CmisObject:
if not title:
title = self.title
new_obj = self.__class__(title=title, name=title, parent=parent)
state = vars(self)
for k, v in state.items():
if not k.startswith("_") and k not in [
"uid",
"id",
"parent",
"title",
"name",
"path",
"subfolders",
"documents",
]:
setattr(new_obj, k, v)
if self.parent:
assert self in self.parent.children
return new_obj
@property
def path(self) -> str:
if self.parent:
return f"{self.parent.path}/{self.title}"
else:
return ""
@property
def is_folder(self) -> bool:
return self.sbe_type == "cmis:folder"
@property
def is_document(self) -> bool:
return self.sbe_type == "cmis:document"
@property
def is_root_folder(self) -> bool:
return self._parent_id is None
@property
def community(self) -> Community | None:
if not self.is_folder:
return self.parent and self.parent.community
if self._community:
return self._community
return self.parent and self.parent.community
@listens_for(CmisObject.name, "set", propagate=True, active_history=True)
def _cmis_sync_name_title(
entity: CmisObject, new_value: str, old_value: symbol | str, initiator: Event
) -> str:
"""Synchronize CmisObject name -> title.
CmisObject.title -> name is done via hybrid_property, avoiding
infinite loop (since "set" is received before attribute has received
value)
"""
if entity.title != new_value:
entity.title = new_value
return new_value
class PathAndSecurityIndexable:
"""Mixin for folder and documents indexation."""
__index_to__ = (
("_indexable_parent_ids", ("parent_ids",)),
("_indexable_roles_and_users", ("allowed_roles_and_users",)),
)
def _iter_to_root(self, skip_self: bool = False) -> Iterator[Document | Folder]:
obj = self if not skip_self else self.parent
while obj:
yield obj
obj = obj.parent
@property
def _indexable_parent_ids(self) -> str:
"""Return a string made of ids separated by a slash: "/1/3/4/5", "5"
being self.parent.id."""
ids = [str(obj.id) for obj in self._iter_to_root(skip_self=True)]
return f"/{'/'.join(reversed(ids))}"
@property
def _indexable_roles_and_users(self) -> str:
"""Returns a string made of type:id elements, like "user:2 group:1
user:6"."""
iter_from_root = reversed(list(self._iter_to_root()))
if self.parent:
# skip root folder only on non-root folder!
next(iter_from_root)
allowed = {o[0] for o in security.get_role_assignements(next(iter_from_root))}
for obj in iter_from_root:
if obj.inherit_security:
continue
obj_allowed = {o[0] for o in security.get_role_assignements(obj)}
if Anonymous in obj_allowed:
continue
parent_allowed = allowed
# pure intersection: users and groups in both are preserved
allowed = allowed & obj_allowed
remaining = parent_allowed - obj_allowed
# find users who can access 'obj' because of their group memberships
# 1. extends groups in obj_allowed with their actual member list
extended_allowed = set(
itertools.chain(
*(p.members if isinstance(p, Group) else (p,) for p in obj_allowed)
)
)
# 2. remaining_users are users explicitly listed in parents but not on
# obj. Are they in a group?
remaining_users = {o for o in remaining if isinstance(o, User)}
allowed |= remaining_users & extended_allowed
# remaining groups: find if some users are eligible
remaining_groups_members = set(
itertools.chain(*(p.members for p in remaining if isinstance(p, Group)))
)
allowed |= remaining_groups_members - extended_allowed
# admin role is always granted access
allowed.add(Admin)
return " ".join(indexable_role(p) for p in allowed)
class Folder(PathAndSecurityIndexable, CmisObject):
__tablename__: str | None = None
sbe_type = "cmis:folder"
__indexable__ = True
_indexable_roles_and_users = PathAndSecurityIndexable._indexable_roles_and_users
parent = relationship(
"Folder",
primaryjoin=(lambda: foreign(CmisObject._parent_id) == remote(Folder.id)),
backref=backref(
"subfolders",
lazy="joined",
order_by="Folder.title",
cascade="all, delete-orphan",
),
)
@property
def icon(self):
return icon_url("folder.png")
@property
def children(self) -> list[Document | Folder]:
return self.subfolders + self.documents
@property
def document_count(self) -> int:
count = len(self.documents)
for f in self.subfolders:
count += f.document_count
return count
@property
def depth(self) -> int:
if self.parent is None:
return 0
return self.parent.depth + 1
def create_subfolder(self, title: str) -> Folder:
subfolder = Folder(title=title, parent=self)
assert subfolder in self.children
return subfolder
def create_document(self, title: str) -> Document:
doc = Document(title=title, parent=self)
assert doc.parent == self
assert doc in self.children
return doc
def get_object_by_path(self, path: str) -> Document | Folder | None:
assert path.startswith("/")
assert "//" not in path
if path == "/":
return self
path_segments = path[1:].split("/")
obj = self
try:
for name in path_segments[:]:
obj = first(x for x in obj.children if x.title == name)
return obj
except IndexError:
return None
def __repr__(self):
return "<{}.{} id={!r} name={!r} path={!r} at 0x{:x}>".format(
self.__class__.__module__,
self.__class__.__name__,
self.id,
self.title,
self.path,
id(self),
)
#
# Security related methods
#
@property
def filtered_children(self) -> list[Folder | Document]:
return security.filter_with_permission(
current_user, "read", self.children, inherit=True
)
@property
def filtered_subfolders(self) -> list[Folder]:
return security.filter_with_permission(
current_user, "read", self.subfolders, inherit=True
)
def get_local_roles_assignments(self):
local_roles_assignments = security.get_role_assignements(self)
# local_roles_assignments = sorted(local_roles_assignments,
# key=lambda u: (u[0].last_name.lower(),
# u[0].first_name.lower()))
return local_roles_assignments
def get_inherited_roles_assignments(self):
if self.parent.is_root_folder:
return []
roles = self.parent.get_local_roles_assignments()
inherited_roles = (
self.parent.get_inherited_roles_assignments()
if self.parent.inherit_security
else []
)
assignments = set(itertools.chain(roles, inherited_roles))
def key(x):
principal = x[0]
if isinstance(principal, User):
# Defensive programming here, this shouldn't happen actually
last_name = principal.last_name or ""
first_name = principal.first_name or ""
return last_name.lower(), first_name.lower()
elif isinstance(principal, Group):
return principal.name
else:
raise Exception(f"Bad class here: {type(principal)}")
return sorted(assignments, key=key)
def members(self) -> Collection[User]:
local_roles = self.get_local_roles_assignments()
inherited_roles = (
self.get_inherited_roles_assignments() if self.inherit_security else []
)
def _iter_users(roles):
for principal, _user in roles:
if isinstance(principal, User):
yield principal
else:
yield from itertools.chain(principal.members, principal.admins)
members = set(_iter_users(itertools.chain(local_roles, inherited_roles)))
members = sorted(members, key=lambda u: (u.last_name, u.first_name))
return members
class BaseContent(CmisObject):
"""A base class for cmisobject with an attached file."""
__tablename__: str | None = None
_content_id = Column(Integer, db.ForeignKey(Blob.id))
content_blob = relationship(Blob, cascade="all, delete", foreign_keys=[_content_id])
#: md5 digest (BTW: not sure they should be part of the public API).
content_digest = Column(Text)
#: size (in bytes) of the content blob.
content_length = Column(
Integer,
default=0,
nullable=False,
server_default=sa.text("0"),
info={
"searchable": True,
"index_to": (("content_length", wf.NUMERIC(stored=True)),),
},
)
#: MIME type of the content stream.
# TODO: normalize mime type?
content_type = Column(
Text,
default="application/octet-stream",
info={"searchable": True, "index_to": (("content_type", wf.ID(stored=True)),)},
)
@property
def content(self) -> bytes:
return self.content_blob.value
@content.setter
def content(self, value: bytes):
assert isinstance(value, bytes)
self.content_blob = Blob()
self.content_blob.value = value
self.content_length = len(value)
def set_content(self, content: bytes, content_type: str = ""):
assert isinstance(content_type, str)
new_digest = md5(content)
if new_digest == self.content_digest:
return
self.content_digest = new_digest
self.content = content
content_type = self.find_content_type(content_type)
if content_type:
self.content_type = content_type
def find_content_type(self, content_type: str = "") -> str:
"""Find possibly more appropriate content_type for this instance.
If `content_type` is a binary one, try to find a better one
based on content name so that 'xxx.pdf' is not flagged as
binary/octet-stream for example
"""
assert isinstance(content_type, str)
if content_type not in (
"",
"application/octet-stream",
"binary/octet-stream",
"application/binary",
"multipart/octet-stream",
):
return content_type
# missing or generic content type: try to find something more useful to be
# able to do preview/indexing/...
if self.title:
guessed_content_type = mimetypes.guess_type(self.title, strict=False)[0]
if (
guessed_content_type
and guessed_content_type != "application/vnd.ms-office.activeX"
):
# mimetypes got an update: "random.bin" would be guessed as
# 'application/vnd.ms-office.activeX'... not so useful in a document
# repository
content_type = guessed_content_type
return content_type
@property
def icon(self) -> str:
return icon_for(self.content_type)
class Document(BaseContent, PathAndSecurityIndexable):
"""A document, in the CMIS sense."""
__tablename__: str | None = None
__indexable__ = True
__index_to__ = (("text", ("text",)),)
_indexable_roles_and_users = PathAndSecurityIndexable._indexable_roles_and_users
parent = relationship(
"Folder",
primaryjoin=(foreign(CmisObject._parent_id) == remote(Folder.id)),
backref=backref(
"documents",
lazy="joined",
order_by="Document.title",
cascade="all, delete-orphan",
),
)
PREVIEW_SIZE = 700
@property
def preview_size(self) -> int:
return self.PREVIEW_SIZE
def has_preview(self, size: int = 0, index: int = 0) -> bool:
if size == 0:
size = self.PREVIEW_SIZE
return converter.has_image(self.content_digest, self.content_type, index, size)
@property
def digest(self) -> str:
"""Alias for content_digest."""
return self.content_digest
_text_id = Column(Integer, db.ForeignKey(Blob.id), info=NOT_AUDITABLE)
text_blob = relationship(Blob, cascade="all, delete", foreign_keys=[_text_id])
_pdf_id = Column(Integer, db.ForeignKey(Blob.id), info=NOT_AUDITABLE)
pdf_blob = relationship(Blob, cascade="all, delete", foreign_keys=[_pdf_id])
_preview_id = Column(Integer, db.ForeignKey(Blob.id), info=NOT_AUDITABLE)
preview_blob = relationship(Blob, cascade="all, delete", foreign_keys=[_preview_id])
language = Column(
Text, info={"searchable": True, "index_to": [("language", wf.ID(stored=True))]}
)
size = Column(Integer)
page_num = Column(Integer, default=1)
# FIXME: use Entity.meta instead
#: Stores extra metadata as a JSON column
extra_metadata_json = Column(UnicodeText, info={"auditable": False})
sbe_type = "cmis:document"
# antivirus status
def ensure_antivirus_scheduled(self) -> bool:
if not self.antivirus_required:
return True
if current_app.config.get("CELERY_ALWAYS_EAGER", False):
async_conversion(self)
return True
task_id = self.content_blob.meta.get("antivirus_task_id")
if task_id is not None:
res = tasks.process_document.AsyncResult(task_id)
if not res.failed():
# success, or pending or running
return True
# schedule a new task
self.content_blob.meta["antivirus_task_id"] = str(uuid.uuid4())
async_conversion(self)
return False
@property
def antivirus_scanned(self) -> bool:
"""True if antivirus task was run, even if antivirus didn't return a
result."""
return self.content_blob and "antivirus" in self.content_blob.meta
@property
def antivirus_status(self) -> bool | None:
"""
True: antivirus has scanned file: no virus
False: antivirus has scanned file: virus detected
None: antivirus task was run, but antivirus didn't return a result
"""
return self.content_blob and self.content_blob.meta.get("antivirus")
@property
def antivirus_required(self) -> bool:
"""True if antivirus doesn't need to be run."""
required = current_app.config["ANTIVIRUS_CHECK_REQUIRED"]
return required and (
not self.antivirus_scanned or self.antivirus_status is None
)
@property
def antivirus_ok(self) -> bool:
"""True if user can safely access document content."""
required = current_app.config["ANTIVIRUS_CHECK_REQUIRED"]
if required:
return self.antivirus_status is True
return self.antivirus_status is not False
# R/W properties
@BaseContent.content.setter
def content(self, value: bytes):
BaseContent.content.fset(self, value)
self.content_blob.meta["antivirus_task_id"] = str(uuid.uuid4())
self.pdf_blob = None
self.text_blob = None
def set_content(self, content: bytes, content_type: str = ""):
super().set_content(content, content_type)
async_conversion(self)
@property
def pdf(self):
if self.pdf_blob:
assert isinstance(self.pdf_blob.value, bytes)
return self.pdf_blob and self.pdf_blob.value
@pdf.setter
def pdf(self, value: bytes):
assert isinstance(value, bytes)
self.pdf_blob = Blob()
self.pdf_blob.value = value
# `text` is an Unicode value.
@property
def text(self) -> str:
return self.text_blob.value.decode("utf8") if self.text_blob is not None else ""
@text.setter
def text(self, value: str):
assert isinstance(value, str)
self.text_blob = Blob()
self.text_blob.value = value.encode("utf8")
@property
def extra_metadata(self) -> dict[str, Any]:
if not hasattr(self, "_extra_metadata"):
if self._extra_metadata is not None:
self._extra_metadata = json.loads(self.extra_metadata_json)
else:
self._extra_metadata = None
return self._extra_metadata
@extra_metadata.setter
def extra_metadata(self, extra_metadata: dict[str, Any]):
self._extra_metadata = extra_metadata
self.extra_metadata_json = str(json.dumps(extra_metadata))
# TODO: or use SQLAlchemy alias?
@property
def file_name(self) -> str:
return self.title
def __repr__(self):
return "<Document id={!r} name={!r} path={!r} content_length={:d} at 0x{:x}>".format(
self.id, self.title, self.path, self.content_length, id(self)
)
# locking management; used for checkin/checkout - this could be generalized to
# any entity
@property
def lock(self) -> Lock | None:
"""
:returns: either `None` if no lock or current lock is expired; either the
current valid :class:`Lock` instance.
"""
lock = self.meta.setdefault("abilian.sbe.documents", {}).get("lock")
if lock:
lock = Lock(**lock)
if lock.expired:
lock = None
return lock
@lock.setter
def lock(self, user):
"""Allow to do `document.lock = user` to set a lock for user.
If user is None, the lock is released.
"""
if user is None:
del self.lock
return
self.set_lock(user=user)
@lock.deleter
def lock(self):
"""Remove lock, if any.
`del document.lock` can be safely done even if no lock is set.
"""
meta = self.meta.setdefault("abilian.sbe.documents", {})
if "lock" in meta:
del meta["lock"]
self.meta.changed()
def set_lock(self, user=None):
if user is None:
user = current_user
lock = self.lock
if lock and not lock.is_owner(user=user):
raise RuntimeError("This document is already locked by another user")
meta = self.meta.setdefault("abilian.sbe.documents", {})
lock = Lock.new()
meta["lock"] = lock.as_dict()
self.meta.changed()
def icon_for(content_type: str) -> str:
for extension, mime_type in mimetypes.types_map.items():
if mime_type == content_type:
extension = extension[1:]
icon = f"{extension}.png"
if icon_exists(icon):
return icon_url(icon)
return icon_url("bin.png")
# Async conversion
_async_data = threading.local()
def _get_documents_queue() -> list[tuple[Document, str]]:
if not hasattr(_async_data, "documents"):
_async_data.documents = []
return _async_data.documents
def async_conversion(document: Document):
_get_documents_queue().append(
(document, document.content_blob.meta.get("antivirus_task_id"))
)
def _trigger_conversion_tasks(session: Session):
if (
# this commit is not from the application session
session is not db.session()
# inside a sub-transaction: not yet written in DB
or session.transaction.nested
):
return
document_queue = _get_documents_queue()
while document_queue:
doc, task_id = document_queue.pop()
if doc.id:
tasks.process_document.apply_async((doc.id,), task_id=task_id)
def setup_listener():
mark_attr = "__abilian_sa_listening"
if getattr(_trigger_conversion_tasks, mark_attr, False):
return
listen(Session, "after_commit", _trigger_conversion_tasks)
setattr(_trigger_conversion_tasks, mark_attr, True)
|
core | tracklist | from __future__ import annotations
import logging
import random
from collections.abc import Iterable
from typing import TYPE_CHECKING, Optional
from mopidy import exceptions
from mopidy.core import listener
from mopidy.internal import deprecation, validation
from mopidy.internal.models import TracklistState
from mopidy.models import TlTrack, Track
from pykka.typing import proxy_method
logger = logging.getLogger(__name__)
if TYPE_CHECKING:
from mopidy.core.actor import Core
from mopidy.types import Query, TracklistField, Uri
class TracklistController:
def __init__(self, core: Core) -> None:
self.core = core
self._next_tlid: int = 1
self._tl_tracks: list[TlTrack] = []
self._version: int = 0
self._consume: bool = False
self._random: bool = False
self._shuffled: list[TlTrack] = []
self._repeat: bool = False
self._single: bool = False
def get_tl_tracks(self) -> list[TlTrack]:
"""Get tracklist as list of :class:`mopidy.models.TlTrack`."""
return self._tl_tracks[:]
def get_tracks(self) -> list[Track]:
"""Get tracklist as list of :class:`mopidy.models.Track`."""
return [tl_track.track for tl_track in self._tl_tracks]
def get_length(self) -> int:
"""Get length of the tracklist."""
return len(self._tl_tracks)
def get_version(self) -> int:
"""Get the tracklist version.
Integer which is increased every time the tracklist is changed. Is not
reset before Mopidy is restarted.
"""
return self._version
def _increase_version(self) -> None:
self._version += 1
self.core.playback._on_tracklist_change()
self._trigger_tracklist_changed()
def get_consume(self) -> bool:
"""Get consume mode.
:class:`True`
Tracks are removed from the tracklist when they have been played.
:class:`False`
Tracks are not removed from the tracklist.
"""
return self._consume
def set_consume(self, value: bool) -> None:
"""Set consume mode.
:class:`True`
Tracks are removed from the tracklist when they have been played.
:class:`False`
Tracks are not removed from the tracklist.
"""
validation.check_boolean(value)
if self.get_consume() != value:
self._trigger_options_changed()
self._consume = value
def get_random(self) -> bool:
"""Get random mode.
:class:`True`
Tracks are selected at random from the tracklist.
:class:`False`
Tracks are played in the order of the tracklist.
"""
return self._random
def set_random(self, value: bool) -> None:
"""Set random mode.
:class:`True`
Tracks are selected at random from the tracklist.
:class:`False`
Tracks are played in the order of the tracklist.
"""
validation.check_boolean(value)
if self.get_random() != value:
self._trigger_options_changed()
if value:
self._shuffled = self.get_tl_tracks()
random.shuffle(self._shuffled)
self._random = value
def get_repeat(self) -> bool:
"""Get repeat mode.
:class:`True`
The tracklist is played repeatedly.
:class:`False`
The tracklist is played once.
"""
return self._repeat
def set_repeat(self, value: bool) -> None:
"""Set repeat mode.
To repeat a single track, set both ``repeat`` and ``single``.
:class:`True`
The tracklist is played repeatedly.
:class:`False`
The tracklist is played once.
"""
validation.check_boolean(value)
if self.get_repeat() != value:
self._trigger_options_changed()
self._repeat = value
def get_single(self) -> bool:
"""Get single mode.
:class:`True`
Playback is stopped after current song, unless in ``repeat`` mode.
:class:`False`
Playback continues after current song.
"""
return self._single
def set_single(self, value: bool) -> None:
"""Set single mode.
:class:`True`
Playback is stopped after current song, unless in ``repeat`` mode.
:class:`False`
Playback continues after current song.
"""
validation.check_boolean(value)
if self.get_single() != value:
self._trigger_options_changed()
self._single = value
def index(
self,
tl_track: Optional[TlTrack] = None,
tlid: Optional[int] = None,
) -> Optional[int]:
"""The position of the given track in the tracklist.
If neither *tl_track* or *tlid* is given we return the index of
the currently playing track.
:param tl_track: the track to find the index of
:param tlid: TLID of the track to find the index of
.. versionadded:: 1.1
The *tlid* parameter
"""
if tl_track is not None:
validation.check_instance(tl_track, TlTrack)
if tlid is not None:
validation.check_integer(tlid, min=1)
if tl_track is None and tlid is None:
tl_track = self.core.playback.get_current_tl_track()
if tl_track is not None:
try:
return self._tl_tracks.index(tl_track)
except ValueError:
pass
elif tlid is not None:
for i, tl_track in enumerate(self._tl_tracks):
if tl_track.tlid == tlid:
return i
return None
def get_eot_tlid(self) -> Optional[int]:
"""The TLID of the track that will be played after the current track.
Not necessarily the same TLID as returned by :meth:`get_next_tlid`.
.. versionadded:: 1.1
"""
current_tl_track = self.core.playback.get_current_tl_track()
with deprecation.ignore("core.tracklist.eot_track"):
eot_tl_track = self.eot_track(current_tl_track)
return getattr(eot_tl_track, "tlid", None)
def eot_track(self, tl_track: Optional[TlTrack]) -> Optional[TlTrack]:
"""The track that will be played after the given track.
Not necessarily the same track as :meth:`next_track`.
.. deprecated:: 3.0
Use :meth:`get_eot_tlid` instead.
:param tl_track: the reference track
"""
deprecation.warn("core.tracklist.eot_track")
if tl_track is not None:
validation.check_instance(tl_track, TlTrack)
if self.get_single() and self.get_repeat():
return tl_track
if self.get_single():
return None
# Current difference between next and EOT handling is that EOT needs to
# handle "single", with that out of the way the rest of the logic is
# shared.
return self.next_track(tl_track)
def get_next_tlid(self) -> Optional[int]:
"""The tlid of the track that will be played if calling
:meth:`mopidy.core.PlaybackController.next()`.
For normal playback this is the next track in the tracklist. If repeat
is enabled the next track can loop around the tracklist. When random is
enabled this should be a random track, all tracks should be played once
before the tracklist repeats.
.. versionadded:: 1.1
"""
current_tl_track = self.core.playback.get_current_tl_track()
with deprecation.ignore("core.tracklist.next_track"):
next_tl_track = self.next_track(current_tl_track)
return getattr(next_tl_track, "tlid", None)
def next_track(self, tl_track: Optional[TlTrack]) -> Optional[TlTrack]:
"""The track that will be played if calling
:meth:`mopidy.core.PlaybackController.next()`.
For normal playback this is the next track in the tracklist. If repeat
is enabled the next track can loop around the tracklist. When random is
enabled this should be a random track, all tracks should be played once
before the tracklist repeats.
.. deprecated:: 3.0
Use :meth:`get_next_tlid` instead.
:param tl_track: the reference track
"""
deprecation.warn("core.tracklist.next_track")
if tl_track is not None:
validation.check_instance(tl_track, TlTrack)
if not self._tl_tracks:
return None
if (
self.get_random()
and not self._shuffled
and (self.get_repeat() or not tl_track)
):
logger.debug("Shuffling tracks")
self._shuffled = self._tl_tracks[:]
random.shuffle(self._shuffled)
if self.get_random():
if self._shuffled:
return self._shuffled[0]
return None
next_index = self.index(tl_track)
if next_index is None:
next_index = 0
else:
next_index += 1
if self.get_repeat():
if self.get_consume() and len(self._tl_tracks) == 1:
return None
next_index %= len(self._tl_tracks)
elif next_index >= len(self._tl_tracks):
return None
return self._tl_tracks[next_index]
def get_previous_tlid(self) -> Optional[int]:
"""Returns the TLID of the track that will be played if calling
:meth:`mopidy.core.PlaybackController.previous()`.
For normal playback this is the previous track in the tracklist. If
random and/or consume is enabled it should return the current track
instead.
.. versionadded:: 1.1
"""
current_tl_track = self.core.playback.get_current_tl_track()
with deprecation.ignore("core.tracklist.previous_track"):
previous_tl_track = self.previous_track(current_tl_track)
return getattr(previous_tl_track, "tlid", None)
def previous_track(self, tl_track: Optional[TlTrack]) -> Optional[TlTrack]:
"""Returns the track that will be played if calling
:meth:`mopidy.core.PlaybackController.previous()`.
For normal playback this is the previous track in the tracklist. If
random and/or consume is enabled it should return the current track
instead.
.. deprecated:: 3.0
Use :meth:`get_previous_tlid` instead.
:param tl_track: the reference track
"""
deprecation.warn("core.tracklist.previous_track")
if tl_track is not None:
validation.check_instance(tl_track, TlTrack)
if self.get_repeat() or self.get_consume() or self.get_random():
return tl_track
position = self.index(tl_track)
if position in (None, 0):
return None
# Since we know we are not at zero we have to be somewhere in the range
# 1 - len(tracks) Thus 'position - 1' will always be within the list.
return self._tl_tracks[position - 1]
def add( # noqa: C901
self,
tracks: Optional[Iterable[Track]] = None,
at_position: Optional[int] = None,
uris: Optional[Iterable[Uri]] = None,
) -> list[TlTrack]:
"""Add tracks to the tracklist.
If ``uris`` is given instead of ``tracks``, the URIs are
looked up in the library and the resulting tracks are added to the
tracklist.
If ``at_position`` is given, the tracks are inserted at the given
position in the tracklist. If ``at_position`` is not given, the tracks
are appended to the end of the tracklist.
Triggers the :meth:`mopidy.core.CoreListener.tracklist_changed` event.
:param tracks: tracks to add
:param at_position: position in tracklist to add tracks
:param uris: list of URIs for tracks to add
.. versionadded:: 1.0
The ``uris`` argument.
.. deprecated:: 1.0
The ``tracks`` argument. Use ``uris``.
"""
if sum(o is not None for o in [tracks, uris]) != 1:
raise ValueError('Exactly one of "tracks" or "uris" must be set')
if tracks is not None:
validation.check_instances(tracks, Track)
if uris is not None:
validation.check_uris(uris)
validation.check_integer(at_position or 0)
if tracks:
deprecation.warn("core.tracklist.add:tracks_arg")
if tracks is None:
tracks = []
assert uris is not None
track_map = self.core.library.lookup(uris=uris)
for uri in uris:
tracks.extend(track_map[uri])
tl_tracks = []
max_length = self.core._config["core"]["max_tracklist_length"]
for track in tracks:
if self.get_length() >= max_length:
raise exceptions.TracklistFull(
f"Tracklist may contain at most {max_length:d} tracks."
)
tl_track = TlTrack(self._next_tlid, track)
self._next_tlid += 1
if at_position is not None:
self._tl_tracks.insert(at_position, tl_track)
at_position += 1
else:
self._tl_tracks.append(tl_track)
tl_tracks.append(tl_track)
if tl_tracks:
self._increase_version()
return tl_tracks
def clear(self) -> None:
"""Clear the tracklist.
Triggers the :meth:`mopidy.core.CoreListener.tracklist_changed` event.
"""
self._tl_tracks = []
self._increase_version()
def filter(self, criteria: Query[TracklistField]) -> list[TlTrack]:
"""Filter the tracklist by the given criteria.
Each rule in the criteria consists of a model field and a list of
values to compare it against. If the model field matches any of the
values, it may be returned.
Only tracks that match all the given criteria are returned.
Examples::
# Returns tracks with TLIDs 1, 2, 3, or 4 (tracklist ID)
filter({'tlid': [1, 2, 3, 4]})
# Returns track with URIs 'xyz' or 'abc'
filter({'uri': ['xyz', 'abc']})
# Returns track with a matching TLIDs (1, 3 or 6) and a
# matching URI ('xyz' or 'abc')
filter({'tlid': [1, 3, 6], 'uri': ['xyz', 'abc']})
:param criteria: one or more rules to match by
"""
tlids = criteria.pop("tlid", [])
validation.check_query(criteria, validation.TRACKLIST_FIELDS.keys())
validation.check_instances(tlids, int)
matches = self._tl_tracks
for key, values in criteria.items():
matches = [ct for ct in matches if getattr(ct.track, key) in values]
if tlids:
matches = [ct for ct in matches if ct.tlid in tlids]
return matches
def move(self, start: int, end: int, to_position: int) -> None:
"""Move the tracks in the slice ``[start:end]`` to ``to_position``.
Triggers the :meth:`mopidy.core.CoreListener.tracklist_changed` event.
:param start: position of first track to move
:param end: position after last track to move
:param to_position: new position for the tracks
"""
if start == end:
end += 1
tl_tracks = self._tl_tracks
# TODO: use validation helpers?
if start >= end:
raise AssertionError("start must be smaller than end")
if start < 0:
raise AssertionError("start must be at least zero")
if end > len(tl_tracks):
raise AssertionError("end can not be larger than tracklist length")
if to_position < 0:
raise AssertionError("to_position must be at least zero")
if to_position > len(tl_tracks):
raise AssertionError("to_position can not be larger than tracklist length")
new_tl_tracks = tl_tracks[:start] + tl_tracks[end:]
for tl_track in tl_tracks[start:end]:
new_tl_tracks.insert(to_position, tl_track)
to_position += 1
self._tl_tracks = new_tl_tracks
self._increase_version()
def remove(self, criteria: Query[TracklistField]) -> list[TlTrack]:
"""Remove the matching tracks from the tracklist.
Uses :meth:`filter()` to lookup the tracks to remove.
Triggers the :meth:`mopidy.core.CoreListener.tracklist_changed` event.
Returns the removed tracks.
:param criteria: one or more rules to match by
"""
tl_tracks = self.filter(criteria)
for tl_track in tl_tracks:
position = self._tl_tracks.index(tl_track)
del self._tl_tracks[position]
self._increase_version()
return tl_tracks
def shuffle(self, start: Optional[int] = None, end: Optional[int] = None) -> None:
"""Shuffles the entire tracklist. If ``start`` and ``end`` is given only
shuffles the slice ``[start:end]``.
Triggers the :meth:`mopidy.core.CoreListener.tracklist_changed` event.
:param start: position of first track to shuffle
:param end: position after last track to shuffle
"""
tl_tracks = self._tl_tracks
# TOOD: use validation helpers?
if start is not None and end is not None and start >= end:
raise AssertionError("start must be smaller than end")
if start is not None and start < 0:
raise AssertionError("start must be at least zero")
if end is not None and end > len(tl_tracks):
raise AssertionError("end can not be larger than tracklist length")
before = tl_tracks[: start or 0]
shuffled = tl_tracks[start:end]
after = tl_tracks[end or len(tl_tracks) :]
random.shuffle(shuffled)
self._tl_tracks = before + shuffled + after
self._increase_version()
def slice(self, start: int, end: int) -> list[TlTrack]:
"""Returns a slice of the tracklist, limited by the given start and end
positions.
:param start: position of first track to include in slice
:param end: position after last track to include in slice
"""
# TODO: validate slice?
return self._tl_tracks[start:end]
def _mark_playing(self, tl_track: TlTrack) -> None:
"""Internal method for :class:`mopidy.core.PlaybackController`."""
if self.get_random() and tl_track in self._shuffled:
self._shuffled.remove(tl_track)
def _mark_unplayable(self, tl_track: Optional[TlTrack]) -> None:
"""Internal method for :class:`mopidy.core.PlaybackController`."""
logger.warning(
"Track is not playable: %s",
tl_track.track.uri if tl_track else None,
)
if self.get_consume() and tl_track is not None:
self.remove({"tlid": [tl_track.tlid]})
if self.get_random() and tl_track in self._shuffled:
self._shuffled.remove(tl_track)
def _mark_played(self, tl_track: Optional[TlTrack]) -> bool:
"""Internal method for :class:`mopidy.core.PlaybackController`."""
if self.get_consume() and tl_track is not None:
self.remove({"tlid": [tl_track.tlid]})
return True
return False
def _trigger_tracklist_changed(self) -> None:
if self.get_random():
self._shuffled = self._tl_tracks[:]
random.shuffle(self._shuffled)
else:
self._shuffled = []
logger.debug("Triggering event: tracklist_changed()")
listener.CoreListener.send("tracklist_changed")
def _trigger_options_changed(self) -> None:
logger.debug("Triggering options changed event")
listener.CoreListener.send("options_changed")
def _save_state(self) -> TracklistState:
return TracklistState(
tl_tracks=self._tl_tracks,
next_tlid=self._next_tlid,
consume=self.get_consume(),
random=self.get_random(),
repeat=self.get_repeat(),
single=self.get_single(),
)
def _load_state(
self,
state: TracklistState,
coverage: Iterable[str],
) -> None:
if state:
if "mode" in coverage:
self.set_consume(state.consume)
self.set_random(state.random)
self.set_repeat(state.repeat)
self.set_single(state.single)
if "tracklist" in coverage:
self._next_tlid = max(state.next_tlid, self._next_tlid)
self._tl_tracks = list(state.tl_tracks)
self._increase_version()
class TracklistControllerProxy:
get_tl_tracks = proxy_method(TracklistController.get_tl_tracks)
get_tracks = proxy_method(TracklistController.get_tracks)
get_length = proxy_method(TracklistController.get_length)
get_version = proxy_method(TracklistController.get_version)
get_consume = proxy_method(TracklistController.get_consume)
set_consume = proxy_method(TracklistController.set_consume)
get_random = proxy_method(TracklistController.get_random)
set_random = proxy_method(TracklistController.set_random)
get_repeat = proxy_method(TracklistController.get_repeat)
set_repeat = proxy_method(TracklistController.set_repeat)
get_single = proxy_method(TracklistController.get_single)
set_single = proxy_method(TracklistController.set_single)
index = proxy_method(TracklistController.index)
get_eot_tlid = proxy_method(TracklistController.get_eot_tlid)
eot_track = proxy_method(TracklistController.eot_track)
get_next_tlid = proxy_method(TracklistController.get_next_tlid)
next_track = proxy_method(TracklistController.next_track)
get_previous_tlid = proxy_method(TracklistController.get_previous_tlid)
previous_track = proxy_method(TracklistController.previous_track)
add = proxy_method(TracklistController.add)
clear = proxy_method(TracklistController.clear)
filter = proxy_method(TracklistController.filter)
move = proxy_method(TracklistController.move)
remove = proxy_method(TracklistController.remove)
shuffle = proxy_method(TracklistController.shuffle)
slice = proxy_method(TracklistController.slice)
|
extractor | cinchcast | # coding: utf-8
from __future__ import unicode_literals
from ..utils import unified_strdate, xpath_text
from .common import InfoExtractor
class CinchcastIE(InfoExtractor):
_VALID_URL = (
r"https?://player\.cinchcast\.com/.*?(?:assetId|show_id)=(?P<id>[0-9]+)"
)
_TESTS = [
{
"url": "http://player.cinchcast.com/?show_id=5258197&platformId=1&assetType=single",
"info_dict": {
"id": "5258197",
"ext": "mp3",
"title": "Train Your Brain to Up Your Game with Coach Mandy",
"upload_date": "20130816",
},
},
{
# Actual test is run in generic, look for undergroundwellness
"url": "http://player.cinchcast.com/?platformId=1&assetType=single&assetId=7141703",
"only_matching": True,
},
]
def _real_extract(self, url):
video_id = self._match_id(url)
doc = self._download_xml(
"http://www.blogtalkradio.com/playerasset/mrss?assetType=single&assetId=%s"
% video_id,
video_id,
)
item = doc.find(".//item")
title = xpath_text(item, "./title", fatal=True)
date_str = xpath_text(item, "./{http://developer.longtailvideo.com/trac/}date")
upload_date = unified_strdate(date_str, day_first=False)
# duration is present but wrong
formats = [
{
"format_id": "main",
"url": item.find("./{http://search.yahoo.com/mrss/}content").attrib[
"url"
],
}
]
backup_url = xpath_text(
item, "./{http://developer.longtailvideo.com/trac/}backupContent"
)
if backup_url:
formats.append(
{
"preference": 2, # seems to be more reliable
"format_id": "backup",
"url": backup_url,
}
)
self._sort_formats(formats)
return {
"id": video_id,
"title": title,
"upload_date": upload_date,
"formats": formats,
}
|
serializers | action | from apps.public_api.serializers.webhooks import (
WebhookCreateSerializer,
WebhookTriggerTypeField,
)
from apps.webhooks.models import Webhook
from common.api_helpers.custom_fields import TeamPrimaryKeyRelatedField
from common.api_helpers.utils import CurrentTeamDefault
from rest_framework import serializers
from rest_framework.validators import UniqueTogetherValidator
class ActionCreateSerializer(WebhookCreateSerializer):
team_id = TeamPrimaryKeyRelatedField(
allow_null=True, default=CurrentTeamDefault(), source="team"
)
user = serializers.CharField(
required=False, source="username", allow_null=True, allow_blank=True
)
trigger_type = WebhookTriggerTypeField(required=False)
forward_whole_payload = serializers.BooleanField(
required=False, source="forward_all"
)
class Meta:
model = Webhook
fields = [
"id",
"name",
"is_webhook_enabled",
"organization",
"team_id",
"user",
"data",
"password",
"authorization_header",
"trigger_template",
"headers",
"url",
"forward_whole_payload",
"http_method",
"trigger_type",
"integration_filter",
]
extra_kwargs = {
"name": {"required": True, "allow_null": False, "allow_blank": False},
"url": {"required": True, "allow_null": False, "allow_blank": False},
"data": {"required": False, "allow_null": True, "allow_blank": True},
"password": {"required": False, "allow_null": True, "allow_blank": True},
"authorization_header": {
"required": False,
"allow_null": True,
"allow_blank": True,
},
"trigger_template": {
"required": False,
"allow_null": True,
"allow_blank": True,
},
"headers": {"required": False, "allow_null": True, "allow_blank": True},
"integration_filter": {"required": False, "allow_null": True},
}
validators = [
UniqueTogetherValidator(
queryset=Webhook.objects.all(), fields=["name", "organization"]
)
]
class ActionUpdateSerializer(ActionCreateSerializer):
user = serializers.CharField(required=False, source="username")
trigger_type = WebhookTriggerTypeField(required=False)
forward_whole_payload = serializers.BooleanField(
required=False, source="forward_all"
)
class Meta(ActionCreateSerializer.Meta):
extra_kwargs = {
"name": {"required": False, "allow_null": False, "allow_blank": False},
"is_webhook_enabled": {"required": False, "allow_null": False},
"user": {"required": False, "allow_null": True, "allow_blank": True},
"password": {"required": False, "allow_null": True, "allow_blank": True},
"authorization_header": {
"required": False,
"allow_null": True,
"allow_blank": True,
},
"trigger_template": {
"required": False,
"allow_null": True,
"allow_blank": True,
},
"headers": {"required": False, "allow_null": True, "allow_blank": True},
"url": {"required": False, "allow_null": False, "allow_blank": False},
"data": {"required": False, "allow_null": True, "allow_blank": True},
"forward_whole_payload": {"required": False, "allow_null": False},
"http_method": {
"required": False,
"allow_null": False,
"allow_blank": False,
},
"integration_filter": {"required": False, "allow_null": True},
}
|
Cactus | setup | import os
import shutil
import subprocess
import sys
from distutils.sysconfig import get_python_lib
from setuptools import setup
VERSION = "3.3.3"
SKELETON_FOLDERS = [
"pages",
"plugins",
"static/css",
"static/images",
"static/js",
"templates",
"locale",
]
SKELETON_GLOB = ["skeleton/{0}/*".format(folder) for folder in SKELETON_FOLDERS]
if "uninstall" in sys.argv:
def run(command):
try:
return subprocess.check_output(command, shell=True).strip()
except subprocess.CalledProcessError:
pass
cactusBinPath = run("which cactus")
cactusPackagePath = None
for p in os.listdir(get_python_lib()):
if p.lower().startswith("cactus") and p.lower().endswith(".egg"):
cactusPackagePath = os.path.join(get_python_lib(), p)
if cactusBinPath and os.path.exists(cactusBinPath):
sys.stdout.write("Removing cactus script at %s" % cactusBinPath)
os.unlink(cactusBinPath)
if cactusPackagePath and os.path.isdir(cactusPackagePath):
sys.stdout.write("Removing cactus package at %s" % cactusPackagePath)
shutil.rmtree(cactusPackagePath)
sys.exit()
if "install" in sys.argv or "bdist_egg" in sys.argv:
# Check if we have an old version of cactus installed
p1 = "/usr/local/bin/cactus.py"
p2 = "/usr/local/bin/cactus.pyc"
if os.path.exists(p1) or os.path.exists(p2):
sys.stdout.write(
"Error: you have an old version of Cactus installed, we need to move it:"
)
if os.path.exists(p1):
sys.stderr.write(" sudo rm %s" % p1)
if os.path.exists(p2):
sys.stderr.write(" sudo rm %s" % p2)
sys.exit()
# From Django
def fullsplit(path, result=None):
"""
Split a pathname into components (the opposite of os.path.join)
in a platform-neutral way.
"""
if result is None:
result = []
head, tail = os.path.split(path)
if head == "":
return [tail] + result
if head == path:
return result
return fullsplit(head, [tail] + result)
EXCLUDE_FROM_PACKAGES = [
"cactus.skeleton",
]
def is_package(package_name):
for pkg in EXCLUDE_FROM_PACKAGES:
if package_name.startswith(pkg):
return False
return True
# Compile the list of packages available, because distutils doesn't have
# an easy way to do this.
packages, package_data = [], {}
root_dir = os.path.dirname(__file__)
if root_dir != "":
os.chdir(root_dir)
cactus_dir = "cactus"
for dirpath, dirnames, filenames in os.walk(cactus_dir):
# Ignore PEP 3147 cache dirs and those whose names start with '.'
dirnames[:] = [d for d in dirnames if not d.startswith(".") and d != "__pycache__"]
parts = fullsplit(dirpath)
package_name = ".".join(parts)
if "__init__.py" in filenames and is_package(package_name):
packages.append(package_name)
elif filenames:
relative_path = []
while ".".join(parts) not in packages:
relative_path.append(parts.pop())
relative_path.reverse()
path = os.path.join(*relative_path)
package_files = package_data.setdefault(".".join(parts), [])
package_files.extend([os.path.join(path, f) for f in filenames])
def find_requirements():
# Find all requirements.VERSION.txt files that match (e.g. Python 2.6 matches
# requirements.2.6.txt, requirements.2.txt, and requirments.txt).
v = [str(x) for x in sys.version_info[:2]]
requirements = []
while True:
reqs_file = ".".join(["requirements"] + v + ["txt"])
try:
with open(os.path.join(root_dir, reqs_file)) as f:
requirements.extend(f.readlines())
except IOError:
pass
try:
v.pop()
except IndexError:
break
return requirements
setup(
name="Cactus",
version=VERSION,
description="Static site generation and deployment.",
long_description=open("README.md").read(),
url="http://github.com/koenbok/Cactus",
author="Koen Bok",
author_email="koen@madebysofa.com",
license="BSD",
packages=packages,
package_data=package_data,
entry_points={
"console_scripts": [
"cactus = cactus.cli:cli_entrypoint",
],
},
install_requires=find_requirements(),
extras_require={
"GCS Deployment": ["google-api-python-client"],
"Cloud Files Deployment": ["pyrax"],
"Mac Native FSEvents": ["macfsevents"],
},
tests_require=open(os.path.join(root_dir, "test_requirements.txt")).readlines(),
zip_safe=False,
classifiers=[
"Development Status :: 4 - Beta",
"Environment :: Console",
"Intended Audience :: Developers",
"License :: OSI Approved :: BSD License",
"Natural Language :: English",
"Operating System :: OS Independent",
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.4",
"Topic :: Internet :: WWW/HTTP",
],
)
|
statistics | answerer | from functools import reduce
from operator import mul
from flask_babel import gettext
keywords = ("min", "max", "avg", "sum", "prod")
# required answerer function
# can return a list of results (any result type) for a given query
def answer(query):
parts = query.query.split()
if len(parts) < 2:
return []
try:
args = list(map(float, parts[1:]))
except:
return []
func = parts[0]
answer = None
if func == "min":
answer = min(args)
elif func == "max":
answer = max(args)
elif func == "avg":
answer = sum(args) / len(args)
elif func == "sum":
answer = sum(args)
elif func == "prod":
answer = reduce(mul, args, 1)
if answer is None:
return []
return [{"answer": str(answer)}]
# required answerer function
# returns information about the answerer
def self_info():
return {
"name": gettext("Statistics functions"),
"description": gettext("Compute {functions} of the arguments").format(
functions="/".join(keywords)
),
"examples": ["avg 123 548 2.04 24.2"],
}
|
fighter | projectedChangeState | import wx
from logbook import Logger
from service.fit import Fit
pyfalog = Logger(__name__)
class CalcChangeProjectedFighterStateCommand(wx.Command):
def __init__(self, fitID, position, state):
wx.Command.__init__(self, True, "Change Projected Fighter State")
self.fitID = fitID
self.position = position
self.state = state
self.savedState = None
def Do(self):
pyfalog.debug(
"Doing changing of projected fighter state to {} at position {} for fit {}".format(
self.state, self.position, self.fitID
)
)
fit = Fit.getInstance().getFit(self.fitID)
fighter = fit.projectedFighters[self.position]
self.savedState = fighter.active
if self.state == self.savedState:
return False
fighter.active = self.state
return True
def Undo(self):
pyfalog.debug(
"Undoing changing of projected fighter state to {} at position {} for fit {}".format(
self.state, self.position, self.fitID
)
)
cmd = CalcChangeProjectedFighterStateCommand(
fitID=self.fitID, position=self.position, state=self.savedState
)
return cmd.Do()
|
workflows | s3_batch_export | import asyncio
import contextlib
import datetime as dt
import io
import json
import posixpath
import typing
from dataclasses import dataclass
import aioboto3
from django.conf import settings
from posthog.batch_exports.service import S3BatchExportInputs
from posthog.temporal.workflows.base import PostHogWorkflow
from posthog.temporal.workflows.batch_exports import (
BatchExportTemporaryFile,
CreateBatchExportRunInputs,
UpdateBatchExportRunStatusInputs,
create_export_run,
get_batch_exports_logger,
get_data_interval,
get_results_iterator,
get_rows_count,
update_export_run_status,
)
from posthog.temporal.workflows.clickhouse import get_client
from temporalio import activity, exceptions, workflow
from temporalio.common import RetryPolicy
def get_allowed_template_variables(inputs) -> dict[str, str]:
"""Derive from inputs a dictionary of supported template variables for the S3 key prefix."""
export_datetime = dt.datetime.fromisoformat(inputs.data_interval_end)
return {
"second": f"{export_datetime:%S}",
"minute": f"{export_datetime:%M}",
"hour": f"{export_datetime:%H}",
"day": f"{export_datetime:%d}",
"month": f"{export_datetime:%m}",
"year": f"{export_datetime:%Y}",
"data_interval_start": inputs.data_interval_start,
"data_interval_end": inputs.data_interval_end,
"table": "events",
}
def get_s3_key(inputs) -> str:
"""Return an S3 key given S3InsertInputs."""
template_variables = get_allowed_template_variables(inputs)
key_prefix = inputs.prefix.format(**template_variables)
base_file_name = f"{inputs.data_interval_start}-{inputs.data_interval_end}"
match inputs.compression:
case "gzip":
file_name = base_file_name + ".jsonl.gz"
case "brotli":
file_name = base_file_name + ".jsonl.br"
case _:
file_name = base_file_name + ".jsonl"
key = posixpath.join(key_prefix, file_name)
if posixpath.isabs(key):
# Keys are relative to root dir, so this would add an extra "/"
key = posixpath.relpath(key, "/")
return key
class UploadAlreadyInProgressError(Exception):
"""Exception raised when an S3MultiPartUpload is already in progress."""
def __init__(self, upload_id):
super().__init__(
f"This upload is already in progress with ID: {upload_id}. Instantiate a new object."
)
class NoUploadInProgressError(Exception):
"""Exception raised when there is no S3MultiPartUpload in progress."""
def __init__(self):
super().__init__(
"No multi-part upload is in progress. Call 'create' to start one."
)
class S3MultiPartUploadState(typing.NamedTuple):
upload_id: str
parts: list[dict[str, str | int]]
Part = dict[str, str | int]
class S3MultiPartUpload:
"""An S3 multi-part upload."""
def __init__(
self,
region_name: str,
bucket_name: str,
key: str,
encryption: str | None,
kms_key_id: str | None,
aws_access_key_id: str | None = None,
aws_secret_access_key: str | None = None,
):
self._session = aioboto3.Session()
self.region_name = region_name
self.aws_access_key_id = aws_access_key_id
self.aws_secret_access_key = aws_secret_access_key
self.bucket_name = bucket_name
self.key = key
self.encryption = encryption
self.kms_key_id = kms_key_id
self.upload_id: str | None = None
self.parts: list[Part] = []
def to_state(self) -> S3MultiPartUploadState:
"""Produce state tuple that can be used to resume this S3MultiPartUpload."""
# The second predicate is trivial but required by type-checking.
if self.is_upload_in_progress() is False or self.upload_id is None:
raise NoUploadInProgressError()
return S3MultiPartUploadState(self.upload_id, self.parts)
@property
def part_number(self):
"""Return the current part number."""
return len(self.parts)
def is_upload_in_progress(self) -> bool:
"""Whether this S3MultiPartUpload is in progress or not."""
if self.upload_id is None:
return False
return True
@contextlib.asynccontextmanager
async def s3_client(self):
async with self._session.client(
"s3",
region_name=self.region_name,
aws_access_key_id=self.aws_access_key_id,
aws_secret_access_key=self.aws_secret_access_key,
) as client:
yield client
async def start(self) -> str:
"""Start this S3MultiPartUpload."""
if self.is_upload_in_progress() is True:
raise UploadAlreadyInProgressError(self.upload_id)
optional_kwargs = {}
if self.encryption:
optional_kwargs["ServerSideEncryption"] = self.encryption
if self.kms_key_id:
optional_kwargs["SSEKMSKeyId"] = self.kms_key_id
async with self.s3_client() as s3_client:
multipart_response = await s3_client.create_multipart_upload(
Bucket=self.bucket_name,
Key=self.key,
**optional_kwargs,
)
upload_id: str = multipart_response["UploadId"]
self.upload_id = upload_id
return upload_id
def continue_from_state(self, state: S3MultiPartUploadState):
"""Continue this S3MultiPartUpload from a previous state."""
self.upload_id = state.upload_id
self.parts = state.parts
return self.upload_id
async def complete(self) -> str:
if self.is_upload_in_progress() is False:
raise NoUploadInProgressError()
async with self.s3_client() as s3_client:
response = await s3_client.complete_multipart_upload(
Bucket=self.bucket_name,
Key=self.key,
UploadId=self.upload_id,
MultipartUpload={"Parts": self.parts},
)
self.upload_id = None
self.parts = []
return response["Location"]
async def abort(self):
if self.is_upload_in_progress() is False:
raise NoUploadInProgressError()
async with self.s3_client() as s3_client:
await s3_client.abort_multipart_upload(
Bucket=self.bucket_name,
Key=self.key,
UploadId=self.upload_id,
)
self.upload_id = None
self.parts = []
async def upload_part(self, body: BatchExportTemporaryFile, rewind: bool = True):
next_part_number = self.part_number + 1
if rewind is True:
body.rewind()
# aiohttp is not duck-type friendly and requires a io.IOBase
# We comply with the file-like interface of io.IOBase.
# So we tell mypy to be nice with us.
reader = io.BufferedReader(body) # type: ignore
async with self.s3_client() as s3_client:
response = await s3_client.upload_part(
Bucket=self.bucket_name,
Key=self.key,
PartNumber=next_part_number,
UploadId=self.upload_id,
Body=reader,
)
reader.detach() # BufferedReader closes the file otherwise.
self.parts.append({"PartNumber": next_part_number, "ETag": response["ETag"]})
async def __aenter__(self):
if not self.is_upload_in_progress():
await self.start()
return self
async def __aexit__(self, exc_type, exc_value, traceback) -> bool:
if exc_value is None:
# Succesfully completed the upload
await self.complete()
return True
if exc_type == asyncio.CancelledError:
# Ensure we clean-up the cancelled upload.
await self.abort()
return False
class HeartbeatDetails(typing.NamedTuple):
"""This tuple allows us to enforce a schema on the Heartbeat details.
Attributes:
last_uploaded_part_timestamp: The timestamp of the last part we managed to upload.
upload_state: State to continue a S3MultiPartUpload when activity execution resumes.
"""
last_uploaded_part_timestamp: str
upload_state: S3MultiPartUploadState
@classmethod
def from_activity_details(cls, details):
last_uploaded_part_timestamp = details[0]
upload_state = S3MultiPartUploadState(*details[1])
return HeartbeatDetails(last_uploaded_part_timestamp, upload_state)
@dataclass
class S3InsertInputs:
"""Inputs for S3 exports."""
# TODO: do _not_ store credentials in temporal inputs. It makes it very hard
# to keep track of where credentials are being stored and increases the
# attach surface for credential leaks.
bucket_name: str
region: str
prefix: str
team_id: int
data_interval_start: str
data_interval_end: str
aws_access_key_id: str | None = None
aws_secret_access_key: str | None = None
compression: str | None = None
exclude_events: list[str] | None = None
encryption: str | None = None
kms_key_id: str | None = None
async def initialize_and_resume_multipart_upload(
inputs: S3InsertInputs
) -> tuple[S3MultiPartUpload, str]:
"""Initialize a S3MultiPartUpload and resume it from a hearbeat state if available."""
logger = get_batch_exports_logger(inputs=inputs)
key = get_s3_key(inputs)
s3_upload = S3MultiPartUpload(
bucket_name=inputs.bucket_name,
key=key,
encryption=inputs.encryption,
kms_key_id=inputs.kms_key_id,
region_name=inputs.region,
aws_access_key_id=inputs.aws_access_key_id,
aws_secret_access_key=inputs.aws_secret_access_key,
)
details = activity.info().heartbeat_details
try:
interval_start, upload_state = HeartbeatDetails.from_activity_details(details)
except IndexError:
# This is the error we expect when no details as the sequence will be empty.
interval_start = inputs.data_interval_start
logger.info(
f"Did not receive details from previous activity Excecution. Export will start from the beginning: {interval_start}"
)
except Exception:
# We still start from the beginning, but we make a point to log unexpected errors.
# Ideally, any new exceptions should be added to the previous block after the first time and we will never land here.
interval_start = inputs.data_interval_start
logger.warning(
f"Did not receive details from previous activity Excecution due to an unexpected error. Export will start from the beginning: {interval_start}",
)
else:
logger.info(
f"Received details from previous activity. Export will attempt to resume from: {interval_start}",
)
s3_upload.continue_from_state(upload_state)
if inputs.compression == "brotli":
# Even if we receive details we cannot resume a brotli compressed upload as we have lost the compressor state.
interval_start = inputs.data_interval_start
logger.info(
f"Export will start from the beginning as we are using brotli compression: {interval_start}",
)
await s3_upload.abort()
return s3_upload, interval_start
@activity.defn
async def insert_into_s3_activity(inputs: S3InsertInputs):
"""
Activity streams data from ClickHouse to S3. It currently only creates a
single file per run, and uploads as a multipart upload.
TODO: this implementation currently tries to export as one run, but it could
be a very big date range and time consuming, better to split into multiple
runs, timing out after say 30 seconds or something and upload multiple
files.
"""
logger = get_batch_exports_logger(inputs=inputs)
logger.info(
"Running S3 export batch %s - %s",
inputs.data_interval_start,
inputs.data_interval_end,
)
async with get_client() as client:
if not await client.is_alive():
raise ConnectionError("Cannot establish connection to ClickHouse")
count = await get_rows_count(
client=client,
team_id=inputs.team_id,
interval_start=inputs.data_interval_start,
interval_end=inputs.data_interval_end,
)
if count == 0:
logger.info(
"Nothing to export in batch %s - %s",
inputs.data_interval_start,
inputs.data_interval_end,
)
return
logger.info("BatchExporting %s rows to S3", count)
s3_upload, interval_start = await initialize_and_resume_multipart_upload(inputs)
# Iterate through chunks of results from ClickHouse and push them to S3
# as a multipart upload. The intention here is to keep memory usage low,
# even if the entire results set is large. We receive results from
# ClickHouse, write them to a local file, and then upload the file to S3
# when it reaches 50MB in size.
results_iterator = get_results_iterator(
client=client,
team_id=inputs.team_id,
interval_start=interval_start,
interval_end=inputs.data_interval_end,
exclude_events=inputs.exclude_events,
)
result = None
last_uploaded_part_timestamp = None
async def worker_shutdown_handler():
"""Handle the Worker shutting down by heart-beating our latest status."""
await activity.wait_for_worker_shutdown()
logger.warn(
f"Worker shutting down! Reporting back latest exported part {last_uploaded_part_timestamp}",
)
activity.heartbeat(last_uploaded_part_timestamp, s3_upload.to_state())
asyncio.create_task(worker_shutdown_handler())
async with s3_upload as s3_upload:
with BatchExportTemporaryFile(
compression=inputs.compression
) as local_results_file:
for result in results_iterator:
record = {
"created_at": result["created_at"],
"distinct_id": result["distinct_id"],
"elements_chain": result["elements_chain"],
"event": result["event"],
"inserted_at": result["inserted_at"],
"person_id": result["person_id"],
"person_properties": result["person_properties"],
"properties": result["properties"],
"timestamp": result["timestamp"],
"uuid": result["uuid"],
}
local_results_file.write_records_to_jsonl([record])
if (
local_results_file.tell()
> settings.BATCH_EXPORT_S3_UPLOAD_CHUNK_SIZE_BYTES
):
logger.info(
"Uploading part %s containing %s records with size %s bytes to S3",
s3_upload.part_number + 1,
local_results_file.records_since_last_reset,
local_results_file.bytes_since_last_reset,
)
await s3_upload.upload_part(local_results_file)
last_uploaded_part_timestamp = result["inserted_at"]
activity.heartbeat(
last_uploaded_part_timestamp, s3_upload.to_state()
)
local_results_file.reset()
if local_results_file.tell() > 0 and result is not None:
logger.info(
"Uploading last part %s containing %s records with size %s bytes to S3",
s3_upload.part_number + 1,
local_results_file.records_since_last_reset,
local_results_file.bytes_since_last_reset,
)
await s3_upload.upload_part(local_results_file)
last_uploaded_part_timestamp = result["inserted_at"]
activity.heartbeat(
last_uploaded_part_timestamp, s3_upload.to_state()
)
@workflow.defn(name="s3-export")
class S3BatchExportWorkflow(PostHogWorkflow):
"""A Temporal Workflow to export ClickHouse data into S3.
This Workflow is intended to be executed both manually and by a Temporal Schedule.
When ran by a schedule, `data_interval_end` should be set to `None` so that we will fetch the
end of the interval from the Temporal search attribute `TemporalScheduledStartTime`.
"""
@staticmethod
def parse_inputs(inputs: list[str]) -> S3BatchExportInputs:
"""Parse inputs from the management command CLI."""
loaded = json.loads(inputs[0])
return S3BatchExportInputs(**loaded)
@workflow.run
async def run(self, inputs: S3BatchExportInputs):
"""Workflow implementation to export data to S3 bucket."""
logger = get_batch_exports_logger(inputs=inputs)
data_interval_start, data_interval_end = get_data_interval(
inputs.interval, inputs.data_interval_end
)
logger.info(
"Starting S3 export batch %s - %s", data_interval_start, data_interval_end
)
create_export_run_inputs = CreateBatchExportRunInputs(
team_id=inputs.team_id,
batch_export_id=inputs.batch_export_id,
data_interval_start=data_interval_start.isoformat(),
data_interval_end=data_interval_end.isoformat(),
)
run_id = await workflow.execute_activity(
create_export_run,
create_export_run_inputs,
start_to_close_timeout=dt.timedelta(minutes=5),
retry_policy=RetryPolicy(
initial_interval=dt.timedelta(seconds=10),
maximum_interval=dt.timedelta(seconds=60),
maximum_attempts=0,
non_retryable_error_types=["NotNullViolation", "IntegrityError"],
),
)
update_inputs = UpdateBatchExportRunStatusInputs(id=run_id, status="Completed")
insert_inputs = S3InsertInputs(
bucket_name=inputs.bucket_name,
region=inputs.region,
prefix=inputs.prefix,
team_id=inputs.team_id,
aws_access_key_id=inputs.aws_access_key_id,
aws_secret_access_key=inputs.aws_secret_access_key,
data_interval_start=data_interval_start.isoformat(),
data_interval_end=data_interval_end.isoformat(),
compression=inputs.compression,
exclude_events=inputs.exclude_events,
encryption=inputs.encryption,
kms_key_id=inputs.kms_key_id,
)
try:
await workflow.execute_activity(
insert_into_s3_activity,
insert_inputs,
start_to_close_timeout=dt.timedelta(minutes=20),
heartbeat_timeout=dt.timedelta(minutes=2),
retry_policy=RetryPolicy(
initial_interval=dt.timedelta(seconds=10),
maximum_interval=dt.timedelta(seconds=120),
maximum_attempts=10,
non_retryable_error_types=[
# S3 parameter validation failed.
"ParamValidationError",
# This error usually indicates credentials are incorrect or permissions are missing.
"ClientError",
],
),
)
except exceptions.ActivityError as e:
if isinstance(e.cause, exceptions.CancelledError):
logger.error("S3 BatchExport was cancelled.")
update_inputs.status = "Cancelled"
else:
logger.exception("S3 BatchExport failed.", exc_info=e.cause)
update_inputs.status = "Failed"
update_inputs.latest_error = str(e.cause)
raise
except Exception as e:
logger.exception(
"S3 BatchExport failed with an unexpected error.", exc_info=e
)
update_inputs.status = "Failed"
update_inputs.latest_error = "An unexpected error has ocurred"
raise
else:
logger.info(
"Successfully finished S3 export batch %s - %s",
data_interval_start,
data_interval_end,
)
finally:
await workflow.execute_activity(
update_export_run_status,
update_inputs,
start_to_close_timeout=dt.timedelta(minutes=5),
retry_policy=RetryPolicy(
initial_interval=dt.timedelta(seconds=10),
maximum_interval=dt.timedelta(seconds=60),
maximum_attempts=0,
non_retryable_error_types=["NotNullViolation", "IntegrityError"],
),
)
|
machine | txbolt | # Copyright (c) 2011 Hesky Fisher
# See LICENSE.txt for details.
"Thread-based monitoring of a stenotype machine using the TX Bolt protocol."
import plover.machine.base
# In the TX Bolt protocol, there are four sets of keys grouped in
# order from left to right. Each byte represents all the keys that
# were pressed in that set. The first two bits indicate which set this
# byte represents. The next bits are set if the corresponding key was
# pressed for the stroke.
# 00XXXXXX 01XXXXXX 10XXXXXX 110XXXXX
# HWPKTS UE*OAR GLBPRF #ZDST
# The protocol uses variable length packets of one, two, three or four
# bytes. Only those bytes for which keys were pressed will be
# transmitted. The bytes arrive in order of the sets so it is clear
# when a new stroke starts. Also, if a key is pressed in an earlier
# set in one stroke and then a key is pressed only in a later set then
# there will be a zero byte to indicate that this is a new stroke. So,
# it is reliable to assume that a stroke ended when a lower set is
# seen. Additionally, if there is no activity then the machine will
# send a zero byte every few seconds.
STENO_KEY_CHART = (
"S-",
"T-",
"K-",
"P-",
"W-",
"H-", # 00
"R-",
"A-",
"O-",
"*",
"-E",
"-U", # 01
"-F",
"-R",
"-P",
"-B",
"-L",
"-G", # 10
"-T",
"-S",
"-D",
"-Z",
"#",
) # 11
class TxBolt(plover.machine.base.SerialStenotypeBase):
"""TX Bolt interface.
This class implements the three methods necessary for a standard
stenotype interface: start_capture, stop_capture, and
add_callback.
"""
KEYS_LAYOUT = """
# # # # # # # # # #
S- T- P- H- * -F -P -L -T -D
S- K- W- R- * -R -B -G -S -Z
A- O- -E -U
"""
def __init__(self, params):
super().__init__(params)
self._reset_stroke_state()
def _reset_stroke_state(self):
self._pressed_keys = []
self._last_key_set = 0
def _finish_stroke(self):
steno_keys = self.keymap.keys_to_actions(self._pressed_keys)
if steno_keys:
self._notify(steno_keys)
self._reset_stroke_state()
def run(self):
"""Overrides base class run method. Do not call directly."""
settings = self.serial_port.getSettingsDict()
settings["timeout"] = 0.1 # seconds
self.serial_port.applySettingsDict(settings)
self._ready()
while not self.finished.is_set():
# Grab data from the serial port, or wait for timeout if none available.
raw = self.serial_port.read(max(1, self.serial_port.inWaiting()))
if not raw:
# Timeout, finish the current stroke.
self._finish_stroke()
continue
for byte in raw:
key_set = byte >> 6
if key_set <= self._last_key_set:
# Starting a new stroke, finish previous one.
self._finish_stroke()
self._last_key_set = key_set
for i in range(5 if key_set == 3 else 6):
if (byte >> i) & 1:
key = STENO_KEY_CHART[(key_set * 6) + i]
self._pressed_keys.append(key)
if key_set == 3:
# Last possible set, the stroke is finished.
self._finish_stroke()
|
PyObjCTest | test_cferror | from CoreFoundation import *
from Foundation import *
from PyObjCTools.TestSupport import *
try:
unicode
except NameError:
unicode = str
try:
long
except NameError:
long = int
try:
unichr
except NameError:
unichr = chr
class TestError(TestCase):
@min_os_level("10.5")
def testTypes(self):
try:
NSCFError = objc.lookUpClass("__NSCFError")
except objc.error:
NSCFError = objc.lookUpClass("NSCFError")
self.assertIs(CFErrorRef, NSCFError)
@min_os_level("10.5")
def testTypeID(self):
self.assertIsInstance(CFErrorGetTypeID(), (int, long))
@min_os_level("10.5")
def testCreation(self):
userInfo = {
b"foo".decode("ascii"): b"bar".decode("ascii"),
}
err = CFErrorCreate(None, kCFErrorDomainPOSIX, 42, userInfo)
self.assertIsInstance(err, CFErrorRef)
self.assertResultIsCFRetained(CFErrorCopyUserInfo)
dct = CFErrorCopyUserInfo(err)
self.assertEqual(dct, userInfo)
keys = [b"key1".decode("ascii"), b"key2".decode("ascii")]
values = [b"value1".decode("ascii"), b"value2".decode("ascii")]
keys = [NSString.stringWithString_(v) for v in keys]
values = [NSString.stringWithString_(v) for v in values]
self.assertArgHasType(CFErrorCreateWithUserInfoKeysAndValues, 3, b"n^@")
self.assertArgSizeInArg(CFErrorCreateWithUserInfoKeysAndValues, 3, 5)
self.assertArgHasType(CFErrorCreateWithUserInfoKeysAndValues, 4, b"n^@")
self.assertArgSizeInArg(CFErrorCreateWithUserInfoKeysAndValues, 4, 5)
err = CFErrorCreateWithUserInfoKeysAndValues(
None, kCFErrorDomainPOSIX, 42, keys, values, 2
)
self.assertIsInstance(err, CFErrorRef)
dct = CFErrorCopyUserInfo(err)
self.assertEqual(
dct,
{
b"key1".decode("ascii"): b"value1".decode("ascii"),
b"key2".decode("ascii"): b"value2".decode("ascii"),
},
)
@min_os_level("10.5")
def testInspection(self):
userInfo = {
b"foo".decode("ascii"): b"bar".decode("ascii"),
kCFErrorLocalizedFailureReasonKey: "failure reason",
kCFErrorLocalizedRecoverySuggestionKey: "recovery suggestion",
}
self.assertResultIsCFRetained(CFErrorCreate)
err = CFErrorCreate(None, kCFErrorDomainPOSIX, 42, userInfo)
self.assertIsInstance(err, CFErrorRef)
dom = CFErrorGetDomain(err)
self.assertEqual(dom, kCFErrorDomainPOSIX)
code = CFErrorGetCode(err)
self.assertEqual(code, 42)
self.assertResultIsCFRetained(CFErrorCopyDescription)
v = CFErrorCopyDescription(err)
self.assertIn(
v,
(
b"Operation could not be completed. failure reason".decode("ascii"),
b"The operation couldn%st be completed. failure reason".decode("ascii")
% (unichr(0x2019),),
),
)
self.assertResultIsCFRetained(CFErrorCopyFailureReason)
v = CFErrorCopyFailureReason(err)
self.assertEqual(v, b"failure reason".decode("ascii"))
self.assertResultIsCFRetained(CFErrorCopyRecoverySuggestion)
v = CFErrorCopyRecoverySuggestion(err)
self.assertEqual(v, b"recovery suggestion".decode("ascii"))
@min_os_level("10.5")
def testConstants(self):
self.assertIsInstance(kCFErrorDomainPOSIX, unicode)
self.assertIsInstance(kCFErrorDomainOSStatus, unicode)
self.assertIsInstance(kCFErrorDomainMach, unicode)
self.assertIsInstance(kCFErrorDomainCocoa, unicode)
self.assertIsInstance(kCFErrorLocalizedDescriptionKey, unicode)
self.assertIsInstance(kCFErrorLocalizedFailureReasonKey, unicode)
self.assertIsInstance(kCFErrorLocalizedRecoverySuggestionKey, unicode)
self.assertIsInstance(kCFErrorDescriptionKey, unicode)
self.assertIsInstance(kCFErrorUnderlyingErrorKey, unicode)
@min_os_level("10.7")
def testConstants10_7(self):
self.assertIsInstance(kCFErrorURLKey, unicode)
self.assertIsInstance(kCFErrorFilePathKey, unicode)
if __name__ == "__main__":
main()
|
unihandecode | krcodepoints | # -*- coding: utf-8 -*-
__license__ = "GPL 3"
__copyright__ = "2010 Hiroshi Miura <miurahr@linux.com>"
__docformat__ = "restructuredtext en"
"""
Unicode code point dictionary.
Based on Unicode.org Unihan database.
"""
CODEPOINTS = {
"x34": [
"Qiu ",
"Tian ",
"",
"",
"Kua ",
"Wu ",
"Yin ",
"",
"",
"",
"",
"",
"Si ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ye ",
"",
"",
"",
"",
"",
"Chou ",
"",
"",
"",
"",
"Nuo ",
"",
"",
"Qiu ",
"",
"",
"",
"Xu ",
"Xing ",
"",
"Xiong ",
"Liu ",
"Lin ",
"Xiang ",
"Yong ",
"Xin ",
"Zhen ",
"Dai ",
"Wu ",
"Pan ",
"",
"",
"Ma ",
"Qian ",
"Yi ",
"Zhong ",
"N ",
"Cheng ",
"Fung ",
"",
"",
"",
"Zhuo ",
"Fang ",
"Ao ",
"Wu ",
"Zuo ",
"",
"Zhou ",
"Dong ",
"Su ",
"Yi ",
"Jiong ",
"Wang ",
"Lei ",
"Nao ",
"Zhu ",
"Suk ",
"",
"",
"",
"Xu ",
"",
"",
"Saan ",
"Jie ",
"Die ",
"Nuo ",
"Su ",
"Yi ",
"Long ",
"Ying ",
"Beng ",
"",
"",
"",
"Lan ",
"Miao ",
"Yi ",
"Li ",
"Ji ",
"Yu ",
"Luo ",
"Chai ",
"Nom ",
"",
"",
"Hun ",
"Xu ",
"Hui ",
"Rao ",
"",
"Zhou ",
"Caam ",
"Han ",
"Xi ",
"Tai ",
"Ai ",
"Hui ",
"Jun ",
"Ma ",
"Lue ",
"Tang ",
"Xiao ",
"Tiao ",
"Zha ",
"Yu ",
"Ku ",
"Er ",
"Nang ",
"Qi ",
"Chi ",
"Mu ",
"Han ",
"Tang ",
"Se ",
"Si ",
"Qiong ",
"Lei ",
"Sa ",
"",
"",
"Hui ",
"Pu ",
"Ta ",
"Shu ",
"Zoeng ",
"Ou ",
"Tai ",
"",
"Mian ",
"Wen ",
"Diao ",
"Yu ",
"Mie ",
"Jun ",
"Niao ",
"Xie ",
"You ",
"",
"",
"She ",
"Fung ",
"Lei ",
"Li ",
"Sin ",
"Luo ",
"Sek ",
"Ji ",
"",
"Kwaan ",
"",
"",
"Quan ",
"",
"Cai ",
"Liang ",
"Gu ",
"Mao ",
"Gung ",
"Gua ",
"Sui ",
"Din ",
"",
"Mao ",
"Man ",
"Hyun ",
"Shi ",
"Li ",
"",
"Wang ",
"Kou ",
"Chui ",
"Zhen ",
"Ding ",
"",
"",
"Bing ",
"Huan ",
"Dong ",
"Gong ",
"Cang ",
"",
"Lian ",
"Jiong ",
"Lu ",
"Xing ",
"",
"Nan ",
"Xie ",
"",
"Bi ",
"Jie ",
"Su ",
"Hung ",
"Geoi6",
"Gung ",
"You ",
"Xing ",
"Qi ",
"Phay ",
"Dian ",
"Fu ",
"Luo ",
"Qia ",
"Jie ",
"Tong ",
"Bak ",
"Yan ",
"Ci ",
"Fan ",
"Lang ",
"",
"Fu ",
"He ",
"Diu ",
"Li ",
"Hua ",
"Tou ",
"Pian ",
"Dai ",
"Jun ",
"E ",
"Qie ",
"Yi ",
"Jue ",
"Rui ",
"Jian ",
"Gong ",
"Chi ",
"Chong ",
"Chi ",
"",
],
"x35": [
"Lue ",
"Daang ",
"Lin ",
"Jue ",
"Su ",
"Xiao ",
"Chan ",
"Put ",
"",
"Zhu ",
"Dan ",
"Jian ",
"Zhou ",
"Duo ",
"Xie ",
"Li ",
"Cim ",
"Chi ",
"Xi ",
"Jian ",
"",
"Ji ",
"",
"Fei ",
"Chu ",
"Bang ",
"Kou ",
"",
"Ba ",
"Liang ",
"Kuai ",
"",
"He ",
"Bit ",
"Jue ",
"Lei ",
"Shen ",
"Pi ",
"Yang ",
"Lu ",
"Bei ",
"E ",
"Lu ",
"",
"Coek ",
"Che ",
"Nuo ",
"Suan ",
"Heng ",
"Yu ",
"",
"Gui ",
"Yi ",
"Xian ",
"Gong ",
"Lou ",
"Cam ",
"Le ",
"Shi ",
"Pei ",
"Sun ",
"Yao ",
"Jie ",
"Zou ",
"",
"Que ",
"Yin ",
"Him ",
"Zhi ",
"Jia ",
"Hu ",
"La ",
"Hou ",
"Ke ",
"Bou ",
"Jing ",
"Ai ",
"Deoi ",
"E ",
"Chu ",
"Xie ",
"Chu ",
"Wei ",
"",
"Teng ",
"Huan ",
"Su ",
"You ",
"Caam ",
"Jun ",
"Zhao ",
"Xu ",
"Shi ",
"",
"Caat ",
"Kui ",
"Soeng ",
"He ",
"Gai ",
"Yan ",
"Qiu ",
"Yi ",
"Hua ",
"Si ",
"Fan ",
"Zhang ",
"Dan ",
"Fang ",
"Song ",
"Ao ",
"Fu ",
"Nei ",
"He ",
"You ",
"Hua ",
"Hung ",
"Chen ",
"Guo ",
"Ng ",
"Hua ",
"Li ",
"Fa ",
"Hao ",
"Pou ",
"Dung6",
"Si ",
"",
"",
"Le ",
"Lin ",
"Yi ",
"Hou ",
"Zaau ",
"Xu ",
"Qu ",
"Er ",
"",
"",
"",
"",
"",
"",
"",
"Nei ",
"Wei ",
"Xie ",
"Ti ",
"Hong ",
"Tun ",
"Bo ",
"Nie ",
"Yin ",
"San ",
"",
"",
"",
"",
"",
"Wai ",
"Shou ",
"Ba ",
"Ye ",
"Ji ",
"Tou ",
"Han ",
"Jiong ",
"Dong ",
"Wen ",
"Lu ",
"Sou ",
"Guo ",
"Ling ",
"",
"Tian ",
"Lun ",
"",
"",
"",
"",
"",
"",
"ge ",
"Ye ",
"Shi ",
"Xue ",
"Fen ",
"Chun ",
"Rou ",
"Duo ",
"Ze ",
"E ",
"Xie ",
"Zaau ",
"E ",
"Sheng ",
"Wen ",
"Man ",
"Hu ",
"Ge ",
"Xia ",
"Man ",
"Bi ",
"Ji ",
"Hou ",
"Zhi ",
"",
"Gaa ",
"",
"",
"Bai ",
"Ai ",
"Ci ",
"Hin ",
"Gou ",
"Dan ",
"Bai ",
"Bo ",
"Na ",
"Li ",
"Xiao ",
"Xiu ",
"",
"",
"",
"",
"",
"Dong ",
"Ti ",
"Cu ",
"Kuo ",
"Lao ",
"Zhi ",
"Ai ",
"Xi ",
"",
"Qie ",
"Zaa ",
"Hei ",
"",
"",
"Chu ",
"Ji ",
"Huo ",
"Ta ",
"Yan ",
"Xu ",
"Put ",
"Sai ",
"",
"",
"",
"Go6",
"Ye ",
"Xiang ",
"Heoi ",
"Xia ",
],
"x36": [
"Zuo ",
"Yi ",
"Ci ",
"",
"",
"Xian ",
"Tai ",
"Rong ",
"Yi ",
"Zhi ",
"Yi ",
"Xian ",
"Ju ",
"Ji ",
"Han ",
"",
"Pao ",
"Li ",
"",
"Lan ",
"Can ",
"Han ",
"Yan ",
"",
"",
"Yan ",
"Han ",
"Haam ",
"Chi ",
"Nian ",
"Huo ",
"",
"Bi ",
"Xia ",
"Weng ",
"Xuan ",
"Waan ",
"You ",
"Qin ",
"Xu ",
"Nei ",
"Bi ",
"Hao ",
"Jing ",
"Ao ",
"Ao ",
"",
"",
"Zam ",
"Taan ",
"Ju ",
"Zaau ",
"Zuo ",
"Bu ",
"Jie ",
"Ai ",
"Zang ",
"Ci ",
"Fa ",
"Zaan ",
"",
"",
"",
"Nie ",
"Liu ",
"Mang ",
"Dui ",
"Bong ",
"Bi ",
"Bao ",
"",
"Chu ",
"Han ",
"Tian ",
"Chang ",
"",
"",
"Do ",
"Wai ",
"Fu ",
"Duo ",
"Yu ",
"Ye ",
"Kui ",
"Han ",
"Kuai ",
"",
"Kuai ",
"Ziu ",
"Long ",
"Sing ",
"Bu ",
"Chi ",
"Xie ",
"Nie ",
"Lang ",
"Yi ",
"Zung ",
"Man ",
"Zhang ",
"Xia ",
"Gun ",
"",
"",
"Ji ",
"Liao ",
"Ye ",
"Ji ",
"Yin ",
"",
"Da ",
"Yi ",
"Xie ",
"Hao ",
"Yong ",
"Han ",
"Chan ",
"Tai ",
"Tang ",
"Zhi ",
"Bao ",
"Meng ",
"Gui ",
"Chan ",
"Lei ",
"",
"Xi ",
"",
"Hei ",
"Qiao ",
"Rang ",
"Yun ",
"",
"Long ",
"Fu ",
"Zuk ",
"",
"Gu ",
"Hoi ",
"Diu ",
"Hua ",
"Guo ",
"",
"Gao ",
"Tao ",
"",
"Shan ",
"Lai ",
"Nie ",
"Fu ",
"Gao ",
"Qie ",
"Ban ",
"Gaa ",
"",
"Xi ",
"Xu ",
"Kui ",
"Meng ",
"Chuo ",
"Hiu ",
"Ji ",
"Nu ",
"Xiao ",
"Yi ",
"Yu ",
"Yi ",
"Yan ",
"",
"Ran ",
"Hao ",
"Sha ",
"Gwan ",
"You ",
"Daam ",
"Xin ",
"Bi ",
"Zaau ",
"Dian ",
"O ",
"Bu ",
"Dung ",
"Si ",
"Er ",
"Si ",
"Mao ",
"Yun ",
"Gei ",
"Naau ",
"Qiao ",
"",
"Pao ",
"",
"Ping ",
"Nuo ",
"Jie ",
"Zi ",
"Er ",
"Duo ",
"",
"",
"",
"Duo ",
"",
"",
"Qie ",
"Leoi ",
"Ou ",
"Sou ",
"Can ",
"Dou ",
"Ci ",
"Peng ",
"Yi ",
"So ",
"Zuo ",
"Po ",
"Qie ",
"Tong ",
"Xin ",
"You ",
"Bei ",
"Long ",
"",
"",
"",
"",
"",
"",
"Ta ",
"Lan ",
"Man ",
"Qiang ",
"Zhou ",
"Yan ",
"Sek ",
"Lu ",
"Sik ",
"Sao ",
"Mian ",
"Fan ",
"Rui ",
"Fa ",
"Cha ",
"Nao ",
"Cing ",
"Chou ",
"Gai ",
"Shu ",
"Pian ",
"Aam ",
"Kui ",
"Sha ",
"Saan ",
"Xian ",
"Zhi ",
],
"x37": [
"",
"",
"Fung ",
"Lian ",
"Xun ",
"Xu ",
"Mi ",
"Hui ",
"Mu ",
"Zung ",
"Pang ",
"Yi ",
"Gou ",
"Tang ",
"Qi ",
"Yun ",
"Shu ",
"Fu ",
"Yi ",
"Da ",
"",
"Lian ",
"Cao ",
"Can ",
"Ju ",
"Lu ",
"Su ",
"Nen ",
"Ao ",
"An ",
"Qian ",
"",
"Ceoi ",
"Sung ",
"",
"Ran ",
"Shen ",
"Mai ",
"Han ",
"Yue ",
"Er ",
"Ao ",
"Xian ",
"Ma ",
"",
"",
"Lan ",
"Hei ",
"Yue ",
"Dong ",
"Weng ",
"Huai ",
"Meng ",
"Niao ",
"Wan ",
"Mi ",
"Nie ",
"Qu ",
"Zan ",
"Lian ",
"Zhi ",
"Zi ",
"Hai ",
"Xu ",
"Hao ",
"Xun ",
"Zhi ",
"Fan ",
"Chun ",
"Gou ",
"",
"Chun ",
"Luan ",
"Zhu ",
"Shou ",
"Liao ",
"Jie ",
"Xie ",
"Ding ",
"Jie ",
"Rong ",
"Mang ",
"Geoi ",
"Ge ",
"Yao ",
"Ning ",
"Yi ",
"Lang ",
"Yong ",
"Yin ",
"",
"Su ",
"Sik ",
"Lin ",
"Ya ",
"Mao ",
"Ming ",
"Zui ",
"Yu ",
"Ye ",
"Gou ",
"Mi ",
"Jun ",
"Wen ",
"",
"Hoi ",
"Dian ",
"Long ",
"",
"Xing ",
"Cui ",
"Qiao ",
"Mian ",
"Meng ",
"Qin ",
"",
"Wan ",
"De ",
"Ai ",
"",
"Bian ",
"Nou ",
"Lian ",
"Jin ",
"Zyu ",
"Chui ",
"Zuo ",
"Bo ",
"Fui ",
"Yao ",
"Tui ",
"Ji ",
"Aam ",
"Guo ",
"Ji ",
"Wei ",
"Bui6",
"Zaat ",
"Xu ",
"Nian ",
"Yun ",
"",
"Ba ",
"Zhe ",
"Ju ",
"Wei ",
"Xi ",
"Qi ",
"Yi ",
"Xie ",
"Ci ",
"Qiu ",
"Tun ",
"Niao ",
"Qi ",
"Ji ",
"Teoi ",
"",
"",
"Dian ",
"Lao ",
"Zhan ",
"Zi ",
"Fan ",
"Yin ",
"Cen ",
"Ji ",
"Hui ",
"Zai ",
"Lan ",
"Nao ",
"Ju ",
"Qin ",
"Dai ",
"Shutsu ",
"Jie ",
"Xu ",
"Cung ",
"Yong ",
"Dou ",
"Chi ",
"Tou ",
"Min ",
"Huang ",
"Sui ",
"Ke ",
"Zu ",
"Hao ",
"Cheng ",
"Xue ",
"Ni ",
"Chi ",
"Lian ",
"An ",
"Chi ",
"Si ",
"Xiang ",
"Yang ",
"Hua ",
"Cuo ",
"Qiu ",
"Lao ",
"Fu ",
"Dui ",
"Mang ",
"Lang ",
"Tuo ",
"Han ",
"Mang ",
"Bo ",
"Kwan ",
"Qi ",
"Han ",
"",
"Long ",
"Ban ",
"Tiao ",
"Lao ",
"Qi ",
"Zan ",
"Mi ",
"Pei ",
"Zhan ",
"Xiang ",
"Gang ",
"",
"Qi ",
"",
"Lu ",
"Caam ",
"Yun ",
"E ",
"Quan ",
"Min ",
"Wei ",
"Quan ",
"Shu ",
"Min ",
"Daat6",
"",
"Ming ",
"Yao ",
"Jue ",
"Li ",
"Kuai ",
"Gang ",
"Yuan ",
"Da ",
"Gou ",
"Lao ",
"Lou ",
"Qian ",
"Ao ",
"Biao ",
"Zung ",
"Mang ",
],
"x38": [
"Dao ",
"Caam ",
"Ao ",
"",
"Xi ",
"Fu ",
"Daan ",
"Jiu ",
"Run ",
"Tong ",
"Qu ",
"E ",
"Hei ",
"Ji ",
"Ji ",
"Hua ",
"Jiao ",
"Zui ",
"Biao ",
"Meng ",
"Bai ",
"Wei ",
"Ji ",
"Ao ",
"Yu ",
"Hao ",
"Dui ",
"Wo ",
"Ni ",
"Cuan ",
"",
"Li ",
"Lu ",
"Niao ",
"Hua ",
"Lai ",
"",
"Lu ",
"Fung ",
"Mi ",
"Yu ",
"Fong ",
"Ju ",
"",
"",
"Zhan ",
"Pang ",
"Yi ",
"",
"Ji ",
"Bi ",
"",
"Ren ",
"Fong ",
"Fan ",
"Ge ",
"Ku ",
"Jie ",
"Miao ",
"Hei ",
"Si ",
"Tong ",
"Zyun ",
"Ci ",
"Bi ",
"Kai ",
"Li ",
"Fong ",
"Sun ",
"Nuo ",
"Fong ",
"Ji ",
"Men ",
"Xian ",
"Qia ",
"E ",
"Mao ",
"",
"Saam ",
"Tou ",
"Zin ",
"Qiao ",
"Zeon ",
"Kwaan ",
"Wu ",
"Zing ",
"Chuang ",
"Ti ",
"Lian ",
"Bi ",
"Haat6",
"Mang ",
"Xue ",
"Feng ",
"Lei ",
"Zou ",
"Zheng ",
"Chu ",
"Man ",
"Long ",
"",
"Yin ",
"Baan ",
"Zheng ",
"Qian ",
"Luan ",
"Nie ",
"Yi ",
"",
"Ji ",
"Ji ",
"Zhai ",
"Yu ",
"Jiu ",
"Huan ",
"Di ",
"Lup ",
"Ling ",
"Ji ",
"Ben ",
"Zha ",
"Ci ",
"Dan ",
"Liao ",
"Yi ",
"Zhao ",
"Xian ",
"Chi ",
"Ci ",
"Chi ",
"Yan ",
"Lang ",
"Dou ",
"Long ",
"Chan ",
"",
"Tui ",
"Cha ",
"Ai ",
"Chi ",
"",
"Ying ",
"Cha ",
"Tou ",
"",
"Tui ",
"Cha ",
"Yao ",
"Zong ",
"Zung ",
"Pun ",
"Qiao ",
"Lian ",
"Qin ",
"Lu ",
"Yan ",
"Kong ",
"Sou ",
"Yi ",
"Chan ",
"Jiong ",
"Jiang ",
"",
"Jing ",
"",
"Dong ",
"Go ",
"Juan ",
"Han ",
"Di ",
"Wu ",
"",
"Hong ",
"Tou ",
"Chi ",
"Min ",
"Bi ",
"",
"Xun ",
"Lu ",
"Sai ",
"She ",
"Bi ",
"",
"Bi ",
"",
"Xian ",
"Wei ",
"Bie ",
"Er ",
"Juan ",
"",
"Zhen ",
"Bei ",
"Yi ",
"Yu ",
"Qu ",
"Zan ",
"Mi ",
"Ni ",
"Si ",
"Gung ",
"",
"Daan ",
"Shan ",
"Tai ",
"Mu ",
"Jing ",
"Bian ",
"Rong ",
"Ceng ",
"Can ",
"Ding ",
"",
"",
"",
"Keoi ",
"Di ",
"Tong ",
"Ta ",
"Xing ",
"Sung ",
"Duo ",
"Xi ",
"Tong ",
"",
"Ti ",
"Shan ",
"Jian ",
"Zhi ",
"Wai ",
"Yin ",
"",
"",
"Huan ",
"Zhong ",
"Qi ",
"Zung ",
"",
"Xie ",
"Xie ",
"Ze ",
"Wei ",
"",
"",
"Ta ",
"Zhan ",
"Ning ",
"",
"",
"Sam ",
"Yi ",
"Ren ",
"Shu ",
"Cha ",
"Zhuo ",
],
"x39": [
"",
"Mian ",
"Ji ",
"Fang ",
"Pei ",
"Ai ",
"Fan ",
"Ao ",
"Qin ",
"Qia ",
"Xiao ",
"Fan ",
"Gaam ",
"Qiao ",
"Go ",
"Tong ",
"Tip ",
"You ",
"Gou ",
"Ben ",
"Fu ",
"Chu ",
"Zhu ",
"",
"Chu ",
"Zaan ",
"Hang ",
"Nin ",
"Jue ",
"Cung ",
"Cha ",
"Kong ",
"Lie ",
"Li ",
"Xu ",
"Paang ",
"Yu ",
"Hai ",
"Li ",
"Hou ",
"Gong ",
"Ke ",
"Yuan ",
"De ",
"Hui ",
"Giu ",
"Kuang ",
"Jiong ",
"Zan ",
"Fu ",
"Qie ",
"Bei ",
"Xi ",
"Ci ",
"Pang ",
"Haam ",
"Xi ",
"Qiu ",
"Huang ",
"Nan ",
"",
"Chou ",
"San ",
"Jim ",
"De ",
"De ",
"Te ",
"Men ",
"Ling ",
"Shou ",
"Dian ",
"Can ",
"Die ",
"Che ",
"Peng ",
"Zi ",
"Ju ",
"Ji ",
"Lai ",
"Tian ",
"Yuan ",
"Zaau ",
"Cai ",
"Qi ",
"Yu ",
"Lian ",
"Cung ",
"",
"",
"",
"Yu ",
"Ji ",
"Wei ",
"Mi ",
"Cui ",
"Xie ",
"Xu ",
"Xi ",
"Qiu ",
"Hui ",
"Ging ",
"Yu ",
"Qie ",
"Shun ",
"Chui ",
"Duo ",
"Lou ",
"Deon ",
"Pang ",
"Tai ",
"Zhou ",
"Yin ",
"Sou ",
"Fei ",
"Shen ",
"Yuan ",
"Yi ",
"Hun ",
"Se ",
"Ye ",
"Min ",
"Fen ",
"He ",
"",
"Yin ",
"Ce ",
"Ni ",
"Ao ",
"Feng ",
"Lian ",
"Chang ",
"Chan ",
"Ma ",
"Di ",
"Hu ",
"Lu ",
"",
"Yi ",
"Hua ",
"Caa ",
"Tui ",
"E ",
"Hua ",
"Sun ",
"Ni ",
"Lian ",
"Li ",
"Xian ",
"Yan ",
"Long ",
"Men ",
"Jian ",
"Gik ",
"",
"Bian ",
"Yu ",
"Huo ",
"Miao ",
"Chou ",
"Hai ",
"",
"Le ",
"Jie ",
"Wei ",
"Yi ",
"Huan ",
"He ",
"Can ",
"Lan ",
"Yin ",
"Xie ",
"Zaat ",
"Luo ",
"Ling ",
"Qian ",
"Huo ",
"Cim ",
"Wo ",
"Zoi ",
"",
"Ge ",
"Zyu ",
"Die ",
"Yong ",
"Ji ",
"Ang ",
"Ru ",
"Xi ",
"Shuang ",
"Xu ",
"Yi ",
"Hu ",
"Ji ",
"Qu ",
"Tian ",
"Sau ",
"Qian ",
"Mu ",
"Gaan ",
"Mao ",
"Yin ",
"Gai ",
"Ba ",
"Xian ",
"Mao ",
"Fang ",
"Ya ",
"Gong ",
"Song ",
"Wei ",
"Xue ",
"Gwaa ",
"Guai ",
"Jiu ",
"E ",
"Zi ",
"Cui ",
"Bi ",
"Wa ",
"Hin ",
"Lie ",
"Gaa ",
"",
"Kuai ",
"",
"Hai ",
"Zaan ",
"Zhu ",
"Chong ",
"Xian ",
"Xuan ",
"So ",
"Qiu ",
"Pei ",
"Gui ",
"Er ",
"Gong ",
"Qiong ",
"Bak6",
"Lao ",
"Li ",
"Chen ",
"San ",
"Bo ",
"Wo ",
"Pou ",
"Hang ",
"Duo ",
"Paang ",
"Te ",
"Ta ",
"Zhi ",
"Biao ",
"Gu ",
"Hot ",
"Coeng ",
],
"x3a": [
"Bing ",
"Zhi ",
"Dong ",
"Cheng ",
"Zhao ",
"Nei ",
"Lin ",
"Po ",
"Ji ",
"Min ",
"Wei ",
"Che ",
"Gou ",
"Bong ",
"Ru ",
"Taan ",
"Bu ",
"Zung ",
"Kui ",
"Lao ",
"Han ",
"Ying ",
"Zhi ",
"Jie ",
"Xing ",
"Xie ",
"Xun ",
"Shan ",
"Qian ",
"Xie ",
"Su ",
"Hai ",
"Mi ",
"Hun ",
"Nang ",
"",
"Hui ",
"Na ",
"Song ",
"Ben ",
"Liu ",
"Jie ",
"Huang ",
"Lan ",
"",
"Hu ",
"Dou ",
"Huo ",
"Ge ",
"Yao ",
"Ce ",
"Gui ",
"Jian ",
"Jian ",
"Chou ",
"Jin ",
"Ma ",
"Hui ",
"Men ",
"Can ",
"Lue ",
"Pi ",
"Yang ",
"Ju ",
"Ju ",
"Que ",
"",
"",
"Shai ",
"Cau ",
"Jiu ",
"Hua ",
"Xian ",
"Xie ",
"Syun ",
"Su ",
"Fei ",
"Ce ",
"Ye ",
"",
"",
"",
"Qin ",
"Hui ",
"Tun ",
"Ling ",
"Qiang ",
"Xi ",
"Yi ",
"Nim ",
"Meng ",
"Tuan ",
"Lan ",
"Hao ",
"Ci ",
"Zhai ",
"Piao ",
"Luo ",
"Mi ",
"Buk ",
"Fu ",
"Cam ",
"Xie ",
"Bo ",
"Hui ",
"Qi ",
"Xie ",
"",
"Hei ",
"Bo ",
"Qian ",
"Ban ",
"Jiao ",
"Jue ",
"Kun ",
"Song ",
"Ju ",
"E ",
"Nie ",
"",
"Die ",
"Die ",
"Pei ",
"Gui ",
"Zi ",
"Qi ",
"Chui ",
"Gwaat6",
"Yu ",
"Qin ",
"Faat ",
"Ke ",
"Fu ",
"Gaang ",
"Di ",
"Xian ",
"Gui ",
"He ",
"Qun ",
"Han ",
"Tong ",
"Bo ",
"Shan ",
"Bi ",
"Lu ",
"Ye ",
"Ni ",
"Chuai ",
"San ",
"Diao ",
"Lu ",
"Tou ",
"Lian ",
"Ke ",
"San ",
"Zhen ",
"Chuai ",
"Lian ",
"Mao ",
"Deon ",
"Qian ",
"Ke ",
"Shao ",
"Qiao ",
"Bi ",
"Zaa ",
"Yin ",
"Kaap ",
"Shan ",
"Su ",
"Sa ",
"Rui ",
"Zhuo ",
"Lu ",
"Ling ",
"Cha ",
"Zaai ",
"Huan ",
"",
"",
"Jia ",
"Ban ",
"Hu ",
"Dou ",
"Caam ",
"Lou ",
"",
"Juan ",
"Ke ",
"Suo ",
"Ge ",
"Zhe ",
"Ding ",
"Duan ",
"Zhu ",
"Yan ",
"Pang ",
"Cha ",
"",
"",
"",
"",
"Yi ",
"Zin ",
"",
"You ",
"Gun ",
"Yao ",
"Yao ",
"Shi ",
"Gong ",
"Qi ",
"Gen ",
"Gwong ",
"",
"Hou ",
"Mi ",
"Fu ",
"Hu ",
"Guang ",
"Dan ",
"Dai ",
"Tou ",
"Yan ",
"",
"Dung ",
"Qu ",
"",
"Chang ",
"Ming ",
"Tou ",
"Bao ",
"On ",
"Ceon ",
"Zung ",
"Xian ",
"",
"",
"",
"Mao ",
"Lang ",
"Nan ",
"Pei ",
"Chen ",
"Hau ",
"Fei ",
"Cou ",
"Gei ",
"Qie ",
"Dai ",
"Sik ",
"Kun ",
"Die ",
"Lu ",
"Thung ",
"",
],
"x3b": [
"",
"Caap ",
"Yu ",
"Tai ",
"Chan ",
"Man ",
"Mian ",
"Huan ",
"Wan ",
"Nuan ",
"Huan ",
"Hou ",
"Jing ",
"Bo ",
"Xian ",
"Li ",
"Jin ",
"",
"Mang ",
"Piao ",
"Hao ",
"Yang ",
"",
"Xian ",
"Su ",
"Wei ",
"Che ",
"Kap ",
"Jin ",
"Ceng ",
"He ",
"",
"Shai ",
"Ling ",
"Fui ",
"Dui ",
"Zaap ",
"Pu ",
"Yue ",
"Bo ",
"",
"Hui ",
"Die ",
"Yan ",
"Ju ",
"Jiao ",
"Kuai ",
"Lie ",
"Yu ",
"Ti ",
"Tim ",
"Wu ",
"Hong ",
"Xiao ",
"Hao ",
"",
"Tiu ",
"Zaang ",
"",
"Huang ",
"Fu ",
"",
"",
"Dun ",
"",
"Reng ",
"Jiao ",
"Gong ",
"Xin ",
"",
"",
"Yuan ",
"Jue ",
"Hua ",
"Sik ",
"Bang ",
"Mou ",
"Cat ",
"Gong ",
"Wei ",
"",
"Mei ",
"Si ",
"Bian ",
"Lu ",
"Keoi ",
"",
"",
"He ",
"She ",
"Lu ",
"Pai ",
"Rong ",
"Qiu ",
"Lie ",
"Gong ",
"Xian ",
"Xi ",
"Hing ",
"",
"Niao ",
"",
"",
"",
"Xie ",
"Lei ",
"Fu ",
"Cuan ",
"Zhuo ",
"Fei ",
"Zuo ",
"Die ",
"Ji ",
"He ",
"Ji ",
"",
"Gin ",
"",
"",
"",
"Tu ",
"Xian ",
"Yan ",
"Tang ",
"Ta ",
"Di ",
"Jue ",
"Ang ",
"Han ",
"Yao ",
"Ju ",
"Rui ",
"Bang ",
"Zeoi ",
"Nie ",
"Tian ",
"Nai ",
"",
"",
"You ",
"Mian ",
"Zin ",
"Bui ",
"Nai ",
"Xing ",
"Qi ",
"Zaan ",
"Gen ",
"Tong ",
"Er ",
"Jia ",
"Qin ",
"Mao ",
"E ",
"Li ",
"Chi ",
"Zong ",
"He ",
"Jie ",
"Ji ",
"",
"Guan ",
"Hou ",
"Gai ",
"Cung ",
"Fen ",
"Se ",
"Waat ",
"Ji ",
"Sik ",
"Qiong ",
"He ",
"Zung ",
"Xian ",
"Jie ",
"Hua ",
"Bi ",
"Sing ",
"Caai ",
"Zhen ",
"Sau ",
"Zin ",
"Shi ",
"Gai ",
"Song ",
"Zhi ",
"Ben ",
"",
"",
"",
"Lang ",
"Bi ",
"Xian ",
"Bang ",
"Dai ",
"Cat ",
"Zi ",
"Pi ",
"Chan ",
"Bi ",
"Su ",
"Huo ",
"Hen ",
"Ying ",
"Chuan ",
"Jiang ",
"Nen ",
"Gu ",
"Fang ",
"",
"",
"Ta ",
"Cui ",
"Sai ",
"De ",
"Ran ",
"Kuan ",
"Che ",
"Da ",
"Hu ",
"Cui ",
"Lu ",
"Juan ",
"Lu ",
"Qian ",
"Pao ",
"Zhen ",
"Fan ",
"Li ",
"Cao ",
"Qi ",
"",
"",
"Ti ",
"Ling ",
"Qu ",
"Lian ",
"Lu ",
"Shu ",
"Gong ",
"Zhe ",
"Biao ",
"Jin ",
"Qing ",
"",
"",
"Zong ",
"Pu ",
"Jin ",
"Biao ",
"Jian ",
"Gun ",
"",
"Ban ",
"Zou ",
"Lie ",
],
"x3c": [
"Li ",
"Luo ",
"Shen ",
"Mian ",
"Jian ",
"Di ",
"Bei ",
"Cim ",
"Lian ",
"Zeon ",
"Xun ",
"Pin ",
"Que ",
"Long ",
"Zui ",
"Gou ",
"Jue ",
"San ",
"She ",
"",
"Xie ",
"Hei ",
"Lan ",
"Cu ",
"Yi ",
"Nuo ",
"Li ",
"Yue ",
"",
"Yi ",
"Ci ",
"Ji ",
"Kang ",
"Xie ",
"Hang ",
"Zi ",
"Ke ",
"Hui ",
"Qu ",
"Hai6",
"Hei ",
"Hoi ",
"Wa ",
"Caam ",
"Xun ",
"Haap6",
"Shen ",
"Kou ",
"Qie ",
"Sha ",
"Xu ",
"Ya ",
"Po ",
"Zu ",
"You ",
"Zi ",
"Lian ",
"Jin ",
"Xia ",
"Yi ",
"Qie ",
"Mi ",
"Jiao ",
"Hei ",
"Chi ",
"Shi ",
"Fong ",
"Yin ",
"Mo ",
"Yi ",
"Hei ",
"Se ",
"Jin ",
"Ye ",
"Zaau ",
"Que ",
"Che ",
"Luan ",
"Kwaan ",
"Zheng ",
"",
"",
"Ho ",
"",
"Se ",
"Gwai ",
"Cui ",
"Zaau ",
"An ",
"Xiu ",
"Can ",
"Chuan ",
"Zha ",
"",
"Ji ",
"Bo ",
"Fu ",
"Sing ",
"Lang ",
"Tui ",
"Sek6",
"Ling ",
"E ",
"Wo ",
"Lian ",
"Du ",
"Men ",
"Lan ",
"Wei ",
"Duan ",
"Kuai ",
"Ai ",
"Zai ",
"Hui ",
"Yi ",
"Mo ",
"Zi ",
"Ben ",
"Beng ",
"",
"Bi ",
"Li ",
"Lu ",
"Luo ",
"Hoi ",
"Dan ",
"Goi ",
"Que ",
"Chen ",
"Hung ",
"Cheng ",
"Jiu ",
"Kou ",
"Ji ",
"Ling ",
"Ci ",
"Shao ",
"Kai ",
"Rui ",
"Chuo ",
"Neng ",
"Zi ",
"Lou ",
"Bao ",
"",
"",
"Bao ",
"Rong ",
"Saan ",
"Lei ",
"Siu ",
"Fu ",
"Qu ",
"Saau ",
"Saa ",
"Zhi ",
"Tan ",
"Rong ",
"Zu ",
"Ying ",
"Mao ",
"Nai ",
"Bian ",
"Saau ",
"Seoi ",
"Tang ",
"Han ",
"Zao ",
"Rong ",
"",
"Dang ",
"Pu ",
"Ziu ",
"Tan ",
"",
"Ran ",
"Ning ",
"Lie ",
"Die ",
"Die ",
"Zhong ",
"Siu ",
"Lu ",
"Dan ",
"Kap ",
"Gui ",
"Ji ",
"Ni ",
"Yi ",
"Nian ",
"Yu ",
"Wang ",
"Guo ",
"Ze ",
"Yan ",
"Cui ",
"Xian ",
"Jiao ",
"Shu ",
"Fu ",
"Pei ",
"Ngoet ",
"Zaau ",
"Zaau ",
"Nhop ",
"Bu ",
"Bian ",
"Chi ",
"Sa ",
"Yi ",
"Bian ",
"",
"Dui ",
"Lan ",
"Zi ",
"Chai ",
"Cung ",
"Xuan ",
"Yu ",
"Yu ",
"Zaau ",
"Hong ",
"Cung ",
"",
"Ta ",
"Gwo ",
"",
"",
"Suk6",
"Ju ",
"Xie ",
"Xi ",
"Jian ",
"Tan ",
"Pan ",
"Ta ",
"Xuan ",
"Xian ",
"Niao ",
"Tan ",
"Gaau ",
"",
"",
"",
"Mi ",
"Ji ",
"Gou ",
"Wen ",
"Faa ",
"Wang ",
"You ",
"Ze ",
"Bi ",
"Mi ",
"Goeng ",
"Xie ",
],
"x3d": [
"Fan ",
"Yi ",
"Dam ",
"Lei ",
"Ying ",
"Siu ",
"Jin ",
"She ",
"Yin ",
"Ji ",
"Zyun ",
"Su ",
"",
"",
"",
"Wang ",
"Mian ",
"Su ",
"Yi ",
"Zai ",
"Se ",
"Ji ",
"Luo ",
"Zaau ",
"Mao ",
"Zha ",
"Sui ",
"Zhi ",
"Bian ",
"Li ",
"Caai ",
"",
"",
"",
"",
"",
"",
"Qiao ",
"Guan ",
"",
"Zhen ",
"Zung ",
"Nie ",
"Jun ",
"Xie ",
"Yao ",
"Xie ",
"Zi ",
"Neng ",
"Sam ",
"Si ",
"Long ",
"Chen ",
"Mi ",
"Que ",
"Dam ",
"Na ",
"",
"",
"",
"Su ",
"Xie ",
"Bo ",
"Ding ",
"Cuan ",
"Fong ",
"Chuang ",
"Che ",
"Han ",
"Dan ",
"Hao ",
"",
"",
"",
"Shen ",
"Mi ",
"Chan ",
"Men ",
"Han ",
"Cui ",
"Jue ",
"He ",
"Fei ",
"Shi ",
"Che ",
"Shen ",
"Nu ",
"Fu ",
"Man ",
"Cing ",
"",
"",
"",
"Yi ",
"Chou ",
"Mei ",
"Faat ",
"Bao ",
"Lei ",
"Ke ",
"Dian ",
"Bi ",
"Sui ",
"Ge ",
"Bi ",
"Yi ",
"Xian ",
"Ni ",
"Ying ",
"Zhu ",
"Chun ",
"Feng ",
"Xu ",
"Piao ",
"Wu ",
"Liao ",
"Cang ",
"Zou ",
"Ceoi ",
"Bian ",
"Yao ",
"Huan ",
"Pai ",
"Sou ",
"",
"Dui ",
"Jing ",
"Xi ",
"Bak6",
"Guo ",
"",
"",
"Yan ",
"Xue ",
"Chu ",
"Heng ",
"Ying ",
"",
"",
"",
"Lian ",
"Xian ",
"Huan ",
"Zaan ",
"",
"Lian ",
"Shan ",
"Cang ",
"Bei ",
"Jian ",
"Shu ",
"Fan ",
"Dian ",
"",
"Ba ",
"Yu ",
"Zyun ",
"",
"Nang ",
"Lei ",
"Yi ",
"Dai ",
"",
"Chan ",
"Chao ",
"Gon ",
"Jin ",
"Nen ",
"Pei ",
"",
"",
"Liao ",
"Mei ",
"Jiu ",
"Siu ",
"Liu ",
"Han ",
"",
"Yong ",
"Jin ",
"Chi ",
"Ren ",
"Nong ",
"",
"",
"Hong ",
"Tian ",
"Bung ",
"Oi ",
"Gwaa ",
"Biu ",
"Bo ",
"Qiong ",
"",
"Shu ",
"Cui ",
"Hui ",
"Chao ",
"Dou ",
"Guai ",
"E ",
"Wei ",
"Fen ",
"Tan ",
"",
"Lun ",
"He ",
"Yong ",
"Hui ",
"Nim6",
"Yu ",
"Zong ",
"Yan ",
"Qiu ",
"Zhao ",
"Jiong ",
"Tai ",
"Zin ",
"",
"Bou ",
"",
"Dut ",
"",
"Tui ",
"Lin ",
"Jiong ",
"Zha ",
"Sing ",
"He ",
"Zing ",
"Xu ",
"",
"",
"Hei ",
"Cui ",
"Qing ",
"Mo ",
"Fung ",
"Zou ",
"Beng ",
"Li ",
"",
"",
"Yan ",
"Ge ",
"Mo ",
"Bei ",
"Juan ",
"Die ",
"Shao ",
"",
"Wu ",
"Yan ",
"",
"Jue ",
"Hin ",
],
"x3e": [
"Tai ",
"Han ",
"",
"Dian ",
"Ji ",
"Jie ",
"",
"",
"Ziu ",
"Xie ",
"La ",
"Fan ",
"Huo ",
"Xi ",
"Nie ",
"Mi ",
"Ran ",
"Cuan ",
"Yin ",
"Mi ",
"",
"Jue ",
"Keoi ",
"Tong ",
"Wan ",
"Ze ",
"Li ",
"Shao ",
"Kong ",
"Kan ",
"Ban ",
"Taai ",
"Tiao ",
"Syu ",
"Bei ",
"Ye ",
"Pian ",
"Chan ",
"Hu ",
"Ken ",
"Gaau ",
"An ",
"Chun ",
"Qian ",
"Bei ",
"Baa ",
"Fen ",
"Fo ",
"Tuo ",
"Tuo ",
"Zuo ",
"Ling ",
"",
"Gui ",
"Zin ",
"Shi ",
"Hou ",
"Lie ",
"Saa ",
"Si ",
"Fung ",
"Bei ",
"Ren ",
"Du ",
"Bo ",
"Liang ",
"Ci ",
"Bi ",
"Ji ",
"Zong ",
"Fai ",
"He ",
"Li ",
"Yuan ",
"Yue ",
"Saau ",
"Chan ",
"Di ",
"Lei ",
"Jin ",
"Chong ",
"Si ",
"Pu ",
"Yi ",
"Goeng ",
"Fun ",
"Huan ",
"Tao ",
"Ru ",
"Ying ",
"Ying ",
"Rao ",
"Yin ",
"Shi ",
"Yin ",
"Jue ",
"Tun ",
"Xuan ",
"Gaa ",
"Zung ",
"Qie ",
"Zhu ",
"Ciu ",
"Zoeng ",
"You ",
"",
"Saan ",
"Xi ",
"Shi ",
"Yi ",
"Mo ",
"Huou ",
"",
"Hu ",
"Xiao ",
"Wu ",
"Gang ",
"Jing ",
"Ting ",
"Shi ",
"Ni ",
"Gang ",
"Ta ",
"Waai ",
"Chu ",
"Chan ",
"Piao ",
"Diao ",
"Nao ",
"Nao ",
"Gan ",
"Gou ",
"Yu ",
"Hou ",
"",
"Si ",
"Ci ",
"Hu ",
"Yang ",
"Zung ",
"Xian ",
"Ban ",
"Rong ",
"Lou ",
"Zhao ",
"Can ",
"Liao ",
"Piao ",
"Hai ",
"Fan ",
"Han ",
"Dan ",
"Zhan ",
"",
"Ta ",
"Zhu ",
"Ban ",
"Jian ",
"Yu ",
"Zhuo ",
"You ",
"Li ",
"Kut ",
"Hei ",
"Cim ",
"Chan ",
"Lian ",
"Heo ",
"Si ",
"Jiu ",
"Pu ",
"Qiu ",
"Gong ",
"Zi ",
"Yu ",
"",
"Si ",
"Reng ",
"Niu ",
"Mei ",
"Baat ",
"Jiu ",
"",
"Xu ",
"Ping ",
"Bian ",
"Mao ",
"",
"",
"",
"",
"Yi ",
"You ",
"Gwai ",
"Ping ",
"Kuk ",
"Bao ",
"Hui ",
"",
"",
"",
"Bu ",
"Mang ",
"La ",
"Tu ",
"Wu ",
"Li ",
"Ling ",
"",
"Ji ",
"Jun ",
"Lit6",
"Duo ",
"Jue ",
"Dai ",
"Bei ",
"",
"",
"",
"",
"",
"La ",
"Bian ",
"Sui ",
"Tu ",
"Die ",
"",
"",
"",
"",
"",
"Duo ",
"",
"",
"Sui ",
"Bi ",
"Tu ",
"Se ",
"Can ",
"Tu ",
"Mian ",
"Zeon ",
"Lu ",
"",
"",
"Zhan ",
"Bi ",
"Ji ",
"Cen ",
"Hyun ",
"Li ",
"",
"",
"Sui ",
"Zung ",
"Shu ",
],
"x3f": [
"",
"",
"E ",
"",
"Gei ",
"",
"",
"Qiong ",
"Luo ",
"Yin ",
"Tun ",
"Gu ",
"Yu ",
"Lei ",
"Bei ",
"Nei ",
"Pian ",
"Lian ",
"Qiu ",
"Lian ",
"Waan ",
"Dong ",
"Li ",
"Ding ",
"Wa ",
"Zhou ",
"Gong ",
"Xing ",
"Ang ",
"Fan ",
"Peng ",
"Bai ",
"Tuo ",
"Syu ",
"E ",
"Bai ",
"Qi ",
"Chu ",
"Gong ",
"Tong ",
"Han ",
"Cheng ",
"Jia ",
"Huan ",
"Xing ",
"Dian ",
"Mai ",
"Dong ",
"E ",
"Ruan ",
"Lie ",
"Sheng ",
"Ou ",
"Di ",
"Yu ",
"Chuan ",
"Rong ",
"Hong ",
"Tang ",
"Cong ",
"Piao ",
"Shuang ",
"Lu ",
"Tong ",
"Zheng ",
"Li ",
"Sa ",
"Ban ",
"Si ",
"Dang ",
"Dong ",
"Guai ",
"Yi ",
"Han ",
"Xie ",
"Luo ",
"Liu ",
"Ham ",
"Dan ",
"",
"Cim ",
"Tan ",
"Saang ",
"",
"",
"You ",
"Nan ",
"",
"Gang ",
"Jun ",
"Chi ",
"Kou ",
"Wan ",
"Li ",
"Liu ",
"Lie ",
"Xia ",
"Baai ",
"An ",
"Yu ",
"Ju ",
"Rou ",
"Xun ",
"Zi ",
"Cuo ",
"Can ",
"Zeng ",
"Yong ",
"Fu ",
"Ruan ",
"Sing ",
"Xi ",
"Shu ",
"Jiao ",
"Jiao ",
"Han ",
"Zhang ",
"Zong ",
"",
"Shui ",
"Chen ",
"Fan ",
"Ji ",
"Zi ",
"",
"Gu ",
"Wu ",
"Cui ",
"Qie ",
"Shu ",
"Hoi ",
"Tuo ",
"Du ",
"Si ",
"Ran ",
"Mu ",
"Fu ",
"Ling ",
"Ji ",
"Xiu ",
"Xuan ",
"Nai ",
"At ",
"Jie ",
"Li ",
"Da ",
"Ji ",
"Zyun ",
"Lu ",
"Shen ",
"Li ",
"Lang ",
"Geng ",
"Yin ",
"Se ",
"Qin ",
"Qie ",
"Che ",
"You ",
"Bu ",
"Huang ",
"Que ",
"Lai ",
"Zaam ",
"Hong ",
"Xu ",
"Bang ",
"Ke ",
"Qi ",
"Gwaai ",
"Sheng ",
"Pin ",
"Gaai ",
"Zhou ",
"Huang ",
"Tui ",
"Hu ",
"Bei ",
"",
"",
"Zaa ",
"Ji ",
"Gu ",
"Sai ",
"Gao ",
"Chai ",
"Ma ",
"Zhu ",
"Tui ",
"Tui ",
"Lian ",
"Lang ",
"Baan ",
"",
"Zing ",
"Dai ",
"Ai ",
"Xian ",
"Gwo ",
"Xi ",
"Zung ",
"Tui ",
"Can ",
"Sao ",
"Cim ",
"Jie ",
"Fen ",
"Qun ",
"",
"Yao ",
"Dao ",
"Jia ",
"Lei ",
"Yan ",
"Lu ",
"Tui ",
"Ying ",
"Pi ",
"Luo ",
"Li ",
"Bie ",
"Hoeng ",
"Mao ",
"Bai ",
"huang ",
"Dau ",
"Yao ",
"He ",
"Chun ",
"Hu ",
"Ning ",
"Chou ",
"Li ",
"Tang ",
"Huan ",
"Bi ",
"Baa ",
"Che ",
"Yang ",
"Da ",
"Ao ",
"Xue ",
"Zi ",
"",
"Daap ",
"Ran ",
"Bong ",
"Zao ",
"Wan ",
"Ta ",
"Bao ",
"Gon ",
"Yan ",
"Gaai ",
"Zhu ",
"Ya ",
],
"x40": [
"Fan ",
"You ",
"On ",
"Tui ",
"Meng ",
"She ",
"Jin ",
"Gu ",
"Qi ",
"Qiao ",
"Jiao ",
"Yan ",
"",
"Kan ",
"Mian ",
"Xian ",
"San ",
"Na ",
"Cin ",
"Huan ",
"Niu ",
"Cheng ",
"Tin ",
"Jue ",
"Xi ",
"Qi ",
"Ang ",
"Mei ",
"Gu ",
"",
"Tou ",
"Fan ",
"Qu ",
"Chan ",
"Shun ",
"Bi ",
"Mao ",
"Shuo ",
"Gu ",
"Hong ",
"Huan ",
"Luo ",
"Hang ",
"Jia ",
"Quan ",
"Goi ",
"Mang ",
"Bu ",
"Gu ",
"Fung ",
"Mu ",
"Ai ",
"Ying ",
"Shun ",
"Lang ",
"Jie ",
"Di ",
"Jie ",
"Cau ",
"Pin ",
"Ren ",
"Yan ",
"Du ",
"Di ",
"",
"Lang ",
"Xian ",
"Biu ",
"Xing ",
"Bei ",
"An ",
"Mi ",
"Qi ",
"Qi ",
"Wo ",
"She ",
"Yu ",
"Jia ",
"Cheng ",
"Yao ",
"Ying ",
"Yang ",
"Ji ",
"Jie ",
"Han ",
"Min ",
"Lou ",
"Kai ",
"Yao ",
"Yan ",
"Sun ",
"Gui ",
"Huang ",
"Ying ",
"Sheng ",
"Cha ",
"Lian ",
"",
"Xuan ",
"Chuan ",
"Che ",
"Ni ",
"Qu ",
"Miao ",
"Huo ",
"Yu ",
"Nan ",
"Hu ",
"Ceng ",
"Biu ",
"Qian ",
"She ",
"Jiang ",
"Ao ",
"Mai ",
"Mang ",
"Zhan ",
"Bian ",
"Jiao ",
"Jue ",
"Nong ",
"Bi ",
"Shi ",
"Li ",
"Mo ",
"Lie ",
"Mie ",
"Mo ",
"Xi ",
"Chan ",
"Qu ",
"Jiao ",
"Huo ",
"Zin ",
"Xu ",
"Nang ",
"Tong ",
"Hou ",
"Yu ",
"",
"Cung ",
"Bo ",
"Zuan ",
"Diu ",
"Chuo ",
"Ci ",
"Jie ",
"Kwai ",
"Xing ",
"Hui ",
"Shi ",
"Gwaat6",
"Caam ",
"",
"Yao ",
"Yu ",
"Bang ",
"Jie ",
"Zhe ",
"Gaa ",
"She ",
"Di ",
"Dong ",
"Ci ",
"Fu ",
"Min ",
"Zhen ",
"Zhen ",
"",
"Yan ",
"Diao ",
"Hong ",
"Gong ",
"Diu6",
"Lue ",
"Guai ",
"La ",
"Cui ",
"Fa ",
"Cuo ",
"Yan ",
"Gung ",
"Jie ",
"Gwaai ",
"Guo ",
"Suo ",
"Wan ",
"Zheng ",
"Nie ",
"Diao ",
"Lai ",
"Ta ",
"Cui ",
"Aa ",
"Gun ",
"",
"",
"Dai ",
"",
"Mian ",
"Gaai ",
"Min ",
"Ju ",
"Yu ",
"Zan ",
"Zhao ",
"Ze ",
"Saang ",
"",
"Pan ",
"He ",
"Gou ",
"Hong ",
"Lao ",
"Wu ",
"Chuo ",
"Hang ",
"Lu ",
"Cu ",
"Lian ",
"Zi ",
"Qiao ",
"Shu ",
"",
"xuan ",
"Cen ",
"Zaam ",
"Hui ",
"Su ",
"Chuang ",
"Deon ",
"Long ",
"",
"Nao ",
"Tan ",
"Dan ",
"Wei ",
"Gan ",
"Da ",
"Li ",
"Caat ",
"Xian ",
"Pan ",
"La ",
"Zyu ",
"Niao ",
"Huai ",
"Ying ",
"Xian ",
"Lan ",
"Mo ",
"Ba ",
"",
"Fu ",
"Bi ",
"Fu ",
],
"x41": [
"Huo ",
"Yi ",
"Liu ",
"Zoeng ",
"Zaan ",
"Juan ",
"Huo ",
"Cheng ",
"Dou ",
"E ",
"",
"Yan ",
"Zhui ",
"Du ",
"Qi ",
"Yu ",
"Quan ",
"Huo ",
"Nie ",
"Heng ",
"Ju ",
"She ",
"",
"",
"Peng ",
"Ming ",
"Cao ",
"Lou ",
"Li ",
"Chun ",
"",
"Cui ",
"Shan ",
"Daam ",
"Qi ",
"",
"Lai ",
"Ling ",
"Liao ",
"Reng ",
"Yu ",
"Nao ",
"Chuo ",
"Qi ",
"Yi ",
"Nian ",
"Fu ",
"Jian ",
"Ya ",
"Fong ",
"Chui ",
"Cin ",
"",
"",
"Bi ",
"Dan ",
"Po ",
"Nian ",
"Zhi ",
"Chao ",
"Tian ",
"Tian ",
"Rou ",
"Yi ",
"Lie ",
"An ",
"He ",
"Qiong ",
"Li ",
"Gwai ",
"Zi ",
"Su ",
"Yuan ",
"Ya ",
"Du ",
"Wan ",
"Gyun ",
"Dong ",
"You ",
"Hui ",
"Jian ",
"Rui ",
"Mang ",
"Ju ",
"Zi ",
"Geoi ",
"An ",
"Sui ",
"Lai ",
"Hun ",
"Qiang ",
"Coeng ",
"Duo ",
"Hung ",
"Na ",
"Can ",
"Ti ",
"Xu ",
"Jiu ",
"Huang ",
"Qi ",
"Jie ",
"Mao ",
"Yan ",
"Heoi ",
"Zhi ",
"Tui ",
"",
"Ai ",
"Pang ",
"Cang ",
"Tang ",
"En ",
"Hun ",
"Qi ",
"Chu ",
"Suo ",
"Zhuo ",
"Nou ",
"Tu ",
"Zu ",
"Lou ",
"Miao ",
"Li ",
"Man ",
"Gu ",
"Cen ",
"Hua ",
"Mei ",
"Gou ",
"Lian ",
"Dao ",
"Shan ",
"Ci ",
"",
"",
"Zhi ",
"Ba ",
"Cui ",
"Qiu ",
"",
"Long ",
"Cim ",
"Fei ",
"Guo ",
"Cheng ",
"Jiu ",
"E ",
"Cung ",
"Jue ",
"Hong ",
"Jiao ",
"Cuan ",
"Yao ",
"Tong ",
"Cha ",
"You ",
"Shu ",
"Yao ",
"Ge ",
"Huan ",
"Lang ",
"Jue ",
"Chen ",
"Cyun ",
"Cyun ",
"Shen ",
"Fo ",
"Ming ",
"Ming ",
"Hung ",
"Chuang ",
"Yun ",
"Han6",
"Jin ",
"Chuo ",
"Zyu ",
"Tan ",
"Hong ",
"Qiong ",
"",
"Cheng ",
"Zaau ",
"Yu ",
"Cheng ",
"Tong ",
"Pun ",
"Qiao ",
"Fo ",
"Ju ",
"Lan ",
"Yi ",
"Rong ",
"Si ",
"Hin ",
"Si ",
"Ngat ",
"Fa ",
"",
"Meng ",
"Gui ",
"",
"",
"Hai ",
"Qiao ",
"Chuo ",
"Que ",
"Dui ",
"Li ",
"Ba ",
"Jie ",
"Seoi ",
"Luo ",
"Deoi ",
"Yun ",
"Zung ",
"Hu ",
"Yin ",
"Pok ",
"Zhi ",
"Lian ",
"Zim ",
"Gan ",
"Jian ",
"Zhou ",
"Zhu ",
"Ku ",
"Na ",
"Dui ",
"Ze ",
"Yang ",
"Zhu ",
"Gong ",
"Yi ",
"Ci ",
"Gei ",
"Chuang ",
"Lao ",
"Ren ",
"Rong ",
"Zing ",
"Na ",
"Ce ",
"Zin ",
"",
"Yi ",
"Jue ",
"Bi ",
"Cheng ",
"Jun ",
"Chou ",
"Hui ",
"Chi ",
"Zhi ",
"Yan ",
"",
],
"x42": [
"Saan ",
"Lun ",
"Bing ",
"Zhao ",
"Han ",
"Yu ",
"Dai ",
"Zhao ",
"Fei ",
"Sha ",
"Ling ",
"Ta ",
"Zeoi ",
"Mang ",
"Ye ",
"Bao ",
"Kui ",
"Gua ",
"Nan ",
"Ge ",
"Gaa ",
"Chi ",
"Fo ",
"Suo ",
"Ci ",
"Zhou ",
"Tai ",
"Kuai ",
"Qin ",
"Seoi ",
"Du ",
"Ce ",
"Huan ",
"Gung ",
"Sai ",
"Zheng ",
"Qian ",
"Gan ",
"Zung ",
"Wei ",
"",
"",
"Xi ",
"Na ",
"Pu ",
"Huai ",
"Ju ",
"Zaan ",
"Sau ",
"Tou ",
"Pan ",
"Ta ",
"Qian ",
"Zung ",
"Rong ",
"Luo ",
"Hu ",
"Sou ",
"Zung ",
"Pu ",
"Mie ",
"Gan ",
"Shuo ",
"Mai ",
"Shu ",
"Ling ",
"Lei ",
"Jiang ",
"Leng ",
"Zhi ",
"Diao ",
"",
"San ",
"Hu ",
"Fan ",
"Mei ",
"Sui ",
"Jian ",
"Tang ",
"Xie ",
"Fu ",
"Mo ",
"Fan ",
"Lei ",
"Can ",
"Ceng ",
"Ling ",
"Zaap ",
"Cong ",
"Yun ",
"Meng ",
"Yu ",
"Zhi ",
"Qi ",
"Dan ",
"Huo ",
"Wei ",
"Tan ",
"Se ",
"Xie ",
"Sou ",
"Song ",
"Cin ",
"Liu ",
"Yi ",
"Aau ",
"Lei ",
"Li ",
"Fei ",
"Lie ",
"Lin ",
"Xian ",
"Yao ",
"Aau ",
"Bie ",
"Xian ",
"Rang ",
"Zhuan ",
"Soeng ",
"Dan ",
"Bian ",
"Ling ",
"Hong ",
"Qi ",
"Liao ",
"Ban ",
"Mi ",
"Hu ",
"Hu ",
"Caap ",
"Ce ",
"Pei ",
"Qiong ",
"Ming ",
"Jiu ",
"Bu ",
"Mei ",
"San ",
"Mei ",
"Zong ",
"",
"Li ",
"Quan ",
"Sam ",
"En ",
"Xiang ",
"Zing ",
"Shi ",
"Zing ",
"Gin ",
"Lan ",
"Huang ",
"Jiu ",
"Yan ",
"Deoi ",
"Sa ",
"Tuan ",
"Xie ",
"Zhe ",
"Men ",
"Xi ",
"Man ",
"Zoeng ",
"Huang ",
"Tan ",
"Xiao ",
"Ya ",
"Bi ",
"Luo ",
"Fan ",
"Li ",
"Cui ",
"Cha ",
"Chou ",
"Di ",
"Kuang ",
"Chu ",
"Cim ",
"Chan ",
"Mi ",
"Qian ",
"Qiu ",
"Zhen ",
"Chai ",
"Heoi ",
"Cim ",
"Gu ",
"Yan ",
"Chi ",
"Guai ",
"Mu ",
"Bo ",
"Kua ",
"Geng ",
"Yao ",
"Mao ",
"Wang ",
"",
"",
"",
"Ru ",
"Jue ",
"Zing ",
"Min ",
"Jiang ",
"O ",
"Zhan ",
"Zuo ",
"Yue ",
"Bing ",
"Nou6",
"Zhou ",
"Bi ",
"Ren ",
"Yu ",
"Gin ",
"Chuo ",
"Er ",
"Yi ",
"Mi ",
"Qing ",
"Zing ",
"Wang ",
"Ji ",
"Bu ",
"Syu ",
"Bie ",
"Fan ",
"Yao ",
"Li ",
"Fan ",
"Qu ",
"Fu ",
"Er ",
"O ",
"Zang ",
"Zim ",
"Huo ",
"Jin ",
"Qi ",
"Ju ",
"Lai ",
"Che ",
"Bei ",
"Niu ",
"Yi ",
"Xu ",
"Liu ",
"Xun ",
"Fu ",
"Cau ",
"Nin ",
"Ting ",
"Beng ",
"Zha ",
"Wui ",
],
"x43": [
"Fo ",
"Zaau ",
"Ou ",
"Shuo ",
"Geng ",
"Tang ",
"Gui ",
"Hui ",
"Ta ",
"Gong ",
"Yao ",
"Daap ",
"Qi ",
"Han ",
"Lue ",
"Mi ",
"Mi ",
"Gin ",
"Lu ",
"Fan ",
"Ou ",
"Mi ",
"Jie ",
"Fu ",
"Mi ",
"Huang ",
"Su ",
"Yao ",
"Nie ",
"Jin ",
"Lian ",
"Bi ",
"Qing ",
"Ti ",
"Ling ",
"Zuan ",
"Zhi ",
"Yin ",
"Dao ",
"Chou ",
"Cai ",
"Mi ",
"Yan ",
"Lan ",
"Chong ",
"Ziu ",
"Soeng ",
"Guan ",
"She ",
"Luo ",
"Fan ",
"Si ",
"Luo ",
"Zhu ",
"Zi ",
"Chou ",
"Juan ",
"Jiong ",
"Er ",
"Yi ",
"Rui ",
"Cai ",
"Ren ",
"Fu ",
"Lan ",
"Sui ",
"Yu ",
"Yao ",
"Dian ",
"Ling ",
"Zhu ",
"Ta ",
"Ping ",
"Qian ",
"Jue ",
"Chui ",
"Bu ",
"Gu ",
"Cun ",
"",
"Han ",
"Han ",
"Mou ",
"Hu ",
"Hong ",
"Di ",
"Fu ",
"Xuan ",
"Mi ",
"Mei ",
"Lang ",
"Gu ",
"Zhao ",
"Ta ",
"Yu ",
"Zong ",
"Li ",
"Liao ",
"Wu ",
"Lei ",
"Ji ",
"Lei ",
"Li ",
"Zong ",
"Bo ",
"Ang ",
"Kui ",
"Tuo ",
"Ping ",
"Cau ",
"Zhao ",
"Gui ",
"Zaan ",
"Xu ",
"Nai ",
"Chuo ",
"Duo ",
"Kaap ",
"Dong ",
"Gui ",
"Bo ",
"Zin ",
"Huan ",
"Xuan ",
"Can ",
"Li ",
"Tui ",
"Huang ",
"Xue ",
"Hu ",
"Bao ",
"Ran ",
"Tiao ",
"Fu ",
"Liao ",
"Zaau ",
"Yi ",
"Shu ",
"Po ",
"He ",
"Cu ",
"Fu ",
"Na ",
"An ",
"Chao ",
"Lu ",
"Zhan ",
"Ta ",
"Fu ",
"Gwaang ",
"Zang ",
"Qiao ",
"Su ",
"Baan ",
"Guan ",
"",
"Fan ",
"Chu ",
"",
"Er ",
"Er ",
"Nuan ",
"Qi ",
"Si ",
"Chu ",
"",
"Yan ",
"Bang ",
"An ",
"Zi ",
"Ne ",
"Chuang ",
"Ba ",
"Ciu ",
"Ti ",
"Han ",
"Zuo ",
"Ba ",
"Zhe ",
"Wa ",
"Sheng ",
"Bi ",
"Er ",
"Zhu ",
"Wu ",
"Wen ",
"Zhi ",
"Zhou ",
"Lu ",
"Wen ",
"Gun ",
"Qiu ",
"La ",
"Zai ",
"Sou ",
"Mian ",
"Zhi ",
"Qi ",
"Cao ",
"Piao ",
"Lian ",
"Saap ",
"Long ",
"Su ",
"Qi ",
"Yuan ",
"Feng ",
"Heoi ",
"Jue ",
"Di ",
"Pian ",
"Guan ",
"Niu ",
"Ren ",
"Zhen ",
"Gai ",
"Pi ",
"Tan ",
"Chao ",
"Chun ",
"Ho ",
"Chun ",
"Mo ",
"Bie ",
"Qi ",
"Shi ",
"Bi ",
"Jue ",
"Si ",
"Taam ",
"Hua ",
"Na ",
"Hui ",
"Kaap ",
"Er ",
"Caau ",
"Mou ",
"Zyu ",
"Xi ",
"Zhi ",
"Ren ",
"Ju ",
"Die ",
"Zhe ",
"Shao ",
"Meng ",
"Bi ",
"Han ",
"Yu ",
"Xian ",
"Pong ",
"Neng ",
"Can ",
"Bu ",
"Bong ",
"Qi ",
],
"x44": [
"Ji ",
"Niao ",
"Lu ",
"Jiong ",
"Han ",
"Yi ",
"Cai ",
"Chun ",
"Zhi ",
"Zi ",
"Da ",
"Cung ",
"Tian ",
"Zhou ",
"Daai ",
"Chun ",
"Cau ",
"Zhe ",
"Zaa ",
"Rou ",
"Bin ",
"Ji ",
"Yi ",
"Du ",
"Jue ",
"Ge ",
"Ji ",
"Dap ",
"Can ",
"Suo ",
"Ruo ",
"Xiang ",
"Huang ",
"Qi ",
"Zhu ",
"Cuo ",
"Chi ",
"Weng ",
"Haap6",
"Kao ",
"Gu ",
"Kai ",
"Fan ",
"Sung ",
"Cao ",
"Zhi ",
"Chan ",
"Lei ",
"Gaau ",
"Zak6",
"Zhe ",
"Yu ",
"Gui ",
"Huang ",
"Jin ",
"Daan ",
"Guo ",
"Sao ",
"Tan ",
"",
"Xi ",
"Man ",
"Duo ",
"Ao ",
"Pi ",
"Wu ",
"Ai ",
"Meng ",
"Pi ",
"Meng ",
"Yang ",
"Zhi ",
"Bo ",
"Ying ",
"Wei ",
"Nao ",
"Lan ",
"Yan ",
"Chan ",
"Quan ",
"Zhen ",
"Pu ",
"",
"Tai ",
"Fei ",
"Shu ",
"",
"Dang ",
"Cha ",
"Ran ",
"Tian ",
"Chi ",
"Ta ",
"Jia ",
"Shun ",
"Huang ",
"Liao ",
"Caa ",
"Dou ",
"",
"Jin ",
"E ",
"Keoi ",
"Fu ",
"Duo ",
"",
"E ",
"",
"Yao ",
"Di ",
"",
"Di ",
"Bu ",
"Man ",
"Che ",
"Lun ",
"Qi ",
"Mu ",
"Can ",
"",
"Zung ",
"Sau ",
"Fan ",
"You ",
"Saau ",
"Da ",
"",
"Su ",
"Fu ",
"Ji ",
"Jiang ",
"Cao ",
"Bo ",
"Teng ",
"Che ",
"Fu ",
"Bu ",
"Wu ",
"Haai ",
"Yang ",
"Ming ",
"Pang ",
"Mang ",
"Zang ",
"Meng ",
"Cao ",
"Tiao ",
"Kai ",
"Bai ",
"Xiao ",
"Xin ",
"Qi ",
"Seoi ",
"",
"Shao ",
"Heng ",
"Niu ",
"Xiao ",
"Chen ",
"Daan ",
"Fan ",
"Yin ",
"Ang ",
"Ran ",
"Ri ",
"Fa ",
"Fan ",
"Qu ",
"Shi ",
"He ",
"Bian ",
"Dai ",
"Mo ",
"Deng ",
"",
"",
"Hong ",
"Zing ",
"Cha ",
"Duo ",
"You ",
"Hao ",
"Tin ",
"Kut ",
"Xian ",
"Lei ",
"Jin ",
"Qi ",
"",
"Mei ",
"Zi ",
"",
"",
"",
"Yan ",
"Yi ",
"Yin ",
"Qi ",
"Zhe ",
"Xi ",
"Yi ",
"Ye ",
"E ",
"",
"Zhi ",
"Han ",
"Chuo ",
"",
"Chun ",
"Bing ",
"Kuai ",
"Chou ",
"",
"Tuo ",
"Qiong ",
"Cung ",
"Jiu ",
"Gwai ",
"Cu ",
"Fu ",
"",
"Meng ",
"Li ",
"Lie ",
"Ta ",
"Zi ",
"Gu ",
"Liang ",
"Fat ",
"La ",
"Dian ",
"Ci ",
"Aang ",
"",
"",
"Ji ",
"",
"Cha ",
"Mao ",
"Du ",
"Zaan ",
"Chai ",
"Rui ",
"Hen ",
"Ruan ",
"",
"Lai ",
"Xing ",
"Gan ",
"Yi ",
"Mei ",
"",
"He ",
"Ji ",
"So ",
"Han ",
],
"x45": [
"Seoi ",
"Li ",
"Zi ",
"Zu ",
"Yao ",
"Geoi ",
"Li ",
"Qi ",
"Gan ",
"Li ",
"",
"",
"",
"",
"Su ",
"Chou ",
"",
"Xie ",
"Bei ",
"Xu ",
"Jing ",
"Pu ",
"Ling ",
"Xiang ",
"Zuo ",
"Diao ",
"Chun ",
"Qing ",
"Nan ",
"",
"Lu ",
"Chi ",
"Shao ",
"Yu ",
"Hua ",
"Li ",
"",
"Siu ",
"",
"Li ",
"",
"",
"Dui ",
"",
"Yi ",
"Ning ",
"Si ",
"Hu ",
"Fu ",
"Zaap ",
"Cheng ",
"Nan ",
"Ce ",
"Gaan ",
"Ti ",
"Qin ",
"Biao ",
"Sui ",
"Wei ",
"Deon ",
"Se ",
"Ai ",
"E ",
"Jie ",
"Kuan ",
"Fei ",
"",
"Yin ",
"Zing ",
"Sao ",
"Dou ",
"Hui ",
"Xie ",
"Ze ",
"Tan ",
"Chang ",
"Zhi ",
"Yi ",
"Fu ",
"E ",
"",
"Jun ",
"Gaa ",
"Cha ",
"Xian ",
"Man ",
"Syun ",
"Bi ",
"Ling ",
"Jie ",
"Kui ",
"Jia ",
"",
"Sang ",
"Lang ",
"Guk ",
"Fei ",
"Lu ",
"Zha ",
"He ",
"",
"Ni ",
"Ying ",
"Xiao ",
"Teng ",
"Lao ",
"Ze ",
"Kui ",
"Goeng ",
"Qian ",
"Ju ",
"Piao ",
"Ban ",
"Dou ",
"Lin ",
"Mi ",
"Zhuo ",
"Xie ",
"Hu ",
"Mi ",
"Gaai ",
"Za ",
"Cong ",
"Ge ",
"Nan ",
"Zhu ",
"Yan ",
"Han ",
"Ceoi ",
"Yi ",
"Luan ",
"Yue ",
"Ran ",
"Ling ",
"Niang ",
"Yu ",
"Nue ",
"Heoi ",
"Yi ",
"Nue ",
"Qin ",
"Qian ",
"Xia ",
"Chu ",
"Jin ",
"Mi ",
"",
"Na ",
"Han ",
"Zu ",
"Xia ",
"Yan ",
"Tu ",
"",
"Wu ",
"Suo ",
"Yin ",
"Chong ",
"Zhou ",
"Mang ",
"Yuan ",
"Nu ",
"Miao ",
"Sao ",
"Wan ",
"Li ",
"",
"Na ",
"Shi ",
"Bi ",
"Ci ",
"Bang ",
"",
"Juan ",
"Xiang ",
"Gui ",
"Pai ",
"Hong ",
"Xun ",
"Zha ",
"Yao ",
"Kwan ",
"",
"He ",
"E ",
"Yang ",
"Tiao ",
"You ",
"Jue ",
"Li ",
"",
"Li ",
"",
"Ji ",
"Hu ",
"Zhan ",
"Fu ",
"Chang ",
"Guan ",
"Ju ",
"Meng ",
"Coeng ",
"Cheng ",
"Mou ",
"Sing ",
"Li ",
"Zaan ",
"",
"Si ",
"Yi ",
"Bing ",
"Cung ",
"Hou ",
"Wan ",
"Chi ",
"",
"Ge ",
"Han ",
"Bo ",
"Saau ",
"Liu ",
"Can ",
"Can ",
"Yi ",
"Xuan ",
"Yan ",
"Suo ",
"Gao ",
"Yong ",
"Zung ",
"Fung ",
"Hong ",
"Yu ",
"Cik ",
"Zhe ",
"Ma ",
"Fung ",
"",
"Shuang ",
"Jin ",
"Guan ",
"Pu ",
"Lin ",
"",
"Ting ",
"Goeng ",
"La ",
"Yi ",
"Zung ",
"Ci ",
"Yan ",
"Jie ",
"Faan ",
"Wei ",
"Xian ",
"Ning ",
],
"x46": [
"Fu ",
"Ge ",
"",
"Mo ",
"Fu ",
"Nai ",
"Xian ",
"Wen ",
"Li ",
"Can ",
"Mie ",
"",
"Ni ",
"Chai ",
"Wan ",
"Xu ",
"Nu ",
"Mai ",
"Co ",
"Kan ",
"Ho ",
"Hang ",
"",
"Faai ",
"Yu ",
"Wei ",
"Zhu ",
"Gei ",
"Gan ",
"Yi ",
"",
"",
"Fu ",
"Bi ",
"Zhu ",
"Zi ",
"Shu ",
"Xia ",
"Ni ",
"",
"Jiao ",
"Xuan ",
"Cung ",
"Nou ",
"Rong ",
"Die ",
"Sa ",
"Sau ",
"",
"Yu ",
"",
"Kaam ",
"Zung ",
"Lu ",
"Han ",
"",
"Yi ",
"Zui ",
"Zhan ",
"Su ",
"Wan ",
"Ni ",
"Guan ",
"Jue ",
"Beng ",
"Can ",
"Zung ",
"Duo ",
"Qi ",
"Yao ",
"Gui ",
"Nuan ",
"Hou ",
"Xun ",
"Xie ",
"",
"Hui ",
"",
"Xie ",
"Bo ",
"Ke ",
"Ceoi ",
"Xu ",
"Bai ",
"Aau ",
"Chu ",
"Bang ",
"Ti ",
"Chu ",
"Chi ",
"Niao ",
"Guan ",
"Feng ",
"Xie ",
"Dang ",
"Duo ",
"Jue ",
"Hui ",
"Zeng ",
"Sa ",
"Duo ",
"Ling ",
"Meng ",
"Fan ",
"Guo ",
"Meng ",
"Long ",
"",
"Ying ",
"Hin ",
"Guan ",
"Cu ",
"Li ",
"Du ",
"Ceng ",
"E ",
"Sin ",
"Saai ",
"",
"De ",
"De ",
"Jiang ",
"Lian ",
"",
"Shao ",
"Xi ",
"Si ",
"Wei ",
"",
"",
"He ",
"You ",
"Lu ",
"Lai ",
"Ou ",
"Sheng ",
"Juan ",
"Qi ",
"",
"Yun ",
"",
"Qi ",
"Zong ",
"Leng ",
"Ji ",
"Mai ",
"Chuang ",
"Nian ",
"Baan ",
"Li ",
"Ling ",
"Gong ",
"Chen ",
"",
"Xian ",
"Hu ",
"Bei ",
"Zu ",
"Dai ",
"Dai ",
"Hun ",
"Soi ",
"Che ",
"Ti ",
"",
"Nuo ",
"Zhi ",
"Liu ",
"Fei ",
"Jiao ",
"Gwaan ",
"Ao ",
"Lin ",
"",
"Reng ",
"Tao ",
"Pi ",
"Xin ",
"Shan ",
"Xie ",
"Wa ",
"Tao ",
"Tin ",
"Xi ",
"Xie ",
"Pi ",
"Yao ",
"Yao ",
"Nu ",
"Hao ",
"Nin ",
"Yin ",
"Fan ",
"Nan ",
"Chi ",
"Wang ",
"Yuan ",
"Xia ",
"Zhou ",
"Yuan ",
"Shi ",
"Mi ",
"",
"Ge ",
"Pao ",
"Fei ",
"Hu ",
"Ni ",
"Ci ",
"Mi ",
"Bian ",
"Gam ",
"Na ",
"Yu ",
"E ",
"Zhi ",
"Nin ",
"Xu ",
"Lue ",
"Hui ",
"Xun ",
"Nao ",
"Han ",
"Jia ",
"Dou ",
"Hua ",
"Tuk ",
"",
"Cu ",
"Xi ",
"Song ",
"Mi ",
"Xin ",
"Wu ",
"Qiong ",
"Zheng ",
"Chou ",
"Xing ",
"Jiu ",
"Ju ",
"Hun ",
"Ti ",
"Man ",
"Jian ",
"Qi ",
"Shou ",
"Lei ",
"Wan ",
"Che ",
"Can ",
"Jie ",
"You ",
"Hui ",
"Zha ",
"Su ",
"Ge ",
],
"x47": [
"Nao ",
"Xi ",
"",
"Deoi ",
"Chi ",
"Wei ",
"Mo ",
"Gun ",
"Cau ",
"",
"Zao ",
"Hui ",
"Luan ",
"Liao ",
"Lao ",
"",
"",
"Qia ",
"Ao ",
"Nie ",
"Sui ",
"Mai ",
"Tan ",
"Xin ",
"Jing ",
"An ",
"Ta ",
"Chan ",
"Wei ",
"Tuan ",
"Ji ",
"Chen ",
"Che ",
"Xu ",
"Xian ",
"Xin ",
"",
"Daan ",
"",
"Nao ",
"",
"Yan ",
"Qiu ",
"Hong ",
"Song ",
"Jun ",
"Liao ",
"Ju ",
"",
"Man ",
"Lie ",
"",
"Chu ",
"Chi ",
"Xiang ",
"Cam ",
"Mei ",
"Shu ",
"Ce ",
"Chi ",
"Gu ",
"Yu ",
"Zaam ",
"",
"Liao ",
"Lao ",
"Shu ",
"Zhe ",
"Soeng ",
"",
"Fat ",
"Fui ",
"E ",
"",
"Sha ",
"Zong ",
"Jue ",
"Jun ",
"",
"Lou ",
"Wei ",
"Cung ",
"Zhu ",
"La ",
"Fun ",
"Zhe ",
"Zhao ",
"Zaau ",
"Yi ",
"",
"Ni ",
"Bo ",
"Syun ",
"Yi ",
"Hao ",
"Ya ",
"Huan ",
"Man ",
"Man ",
"Qu ",
"Lao ",
"Hao ",
"",
"Men ",
"Xian ",
"Zhen ",
"Shu ",
"Zuo ",
"Zhu ",
"Gou ",
"Xuan ",
"Yi ",
"Ti ",
"xie ",
"Jin ",
"Can ",
"Zai ",
"Bu ",
"Liang ",
"Zhi ",
"Ji ",
"Wan ",
"Guan ",
"Ceoi ",
"Qing ",
"Ai ",
"Fu ",
"Gui ",
"Gou ",
"Xian ",
"Ruan ",
"Zhi ",
"Biao ",
"Yi ",
"Suo ",
"Die ",
"Gui ",
"Sheng ",
"Xun ",
"Chen ",
"She ",
"Qing ",
"",
"",
"Chun ",
"Hong ",
"Dong ",
"Cheng ",
"Wei ",
"Die ",
"Shu ",
"Caai ",
"Ji ",
"Za ",
"Qi ",
"",
"Fu ",
"Ao ",
"Fu ",
"Po ",
"",
"Tan ",
"Zha ",
"Che ",
"Qu ",
"You ",
"He ",
"Hou ",
"Gui ",
"E ",
"Jiang ",
"Yun ",
"Tou ",
"Qiu ",
"",
"Fu ",
"Zuo ",
"Hu ",
"",
"Bo ",
"",
"Jue ",
"Di ",
"Jue ",
"Fu ",
"Huang ",
"",
"Yong ",
"Chui ",
"Suo ",
"Chi ",
"Hin ",
"",
"",
"Man ",
"Ca ",
"Qi ",
"Jian ",
"Bi ",
"Gei ",
"Zhi ",
"Zhu ",
"Qu ",
"Zhan ",
"Ji ",
"Dian ",
"",
"Li ",
"Li ",
"La ",
"Quan ",
"Ding ",
"Fu ",
"Cha ",
"Tang ",
"Shi ",
"Hang ",
"Qie ",
"Qi ",
"Bo ",
"Na ",
"Tou ",
"Chu ",
"Cu ",
"Yue ",
"Di ",
"Chen ",
"Chu ",
"Bi ",
"Mang ",
"Ba ",
"Tian ",
"Min ",
"Lie ",
"Feng ",
"Caang ",
"Qiu ",
"Tiao ",
"Fu ",
"Kuo ",
"Jian ",
"Ci ",
"",
"",
"Zhen ",
"Qiu ",
"Cuo ",
"Chi ",
"Kui ",
"Lie ",
"Bang ",
"Du ",
"Wu ",
"So ",
"Jue ",
"Lu ",
],
"x48": [
"Chang ",
"Zai ",
"Chu ",
"Liang ",
"Tian ",
"Kun ",
"Chang ",
"Jue ",
"Tu ",
"Hua ",
"Fei ",
"Bi ",
"",
"Qia ",
"Wo ",
"Ji ",
"Qu ",
"Kui ",
"Hu ",
"Cu ",
"Sui ",
"",
"",
"Qiu ",
"Pi ",
"Bei ",
"Wa ",
"Jiao ",
"Rong ",
"",
"Cu ",
"Die ",
"Chi ",
"Cuo ",
"Meng ",
"Xuan ",
"Duo ",
"Bie ",
"Zhe ",
"Chu ",
"Chan ",
"Gui ",
"Duan ",
"Zou ",
"Deng ",
"Lai ",
"Teng ",
"Yue ",
"Quan ",
"Shu ",
"Ling ",
"",
"Qin ",
"Fu ",
"She ",
"Tiao ",
"",
"Ai ",
"",
"Qiong ",
"Diao ",
"Hai ",
"Shan ",
"Wai ",
"Zhan ",
"Long ",
"Jiu ",
"Li ",
"",
"Min ",
"Rong ",
"Yue ",
"Jue ",
"Kang ",
"Fan ",
"Qi ",
"Hong ",
"Fu ",
"Lu ",
"Hong ",
"Tuo ",
"Min ",
"Tian ",
"Juan ",
"Qi ",
"Zheng ",
"Jing ",
"Gong ",
"Tian ",
"Lang ",
"Mao ",
"Yin ",
"Lu ",
"Yun ",
"Ju ",
"Pi ",
"",
"Xie ",
"Bian ",
"",
"Zeoi ",
"Rong ",
"Sang ",
"Wu ",
"Cha ",
"Gu ",
"Chan ",
"Peng ",
"Man ",
"Saau ",
"Zung ",
"Shuang ",
"Keng ",
"Zhuan ",
"Chan ",
"Si ",
"Chuang ",
"Sui ",
"Bei ",
"Kai ",
"",
"Zhi ",
"Wei ",
"Min ",
"Ling ",
"",
"Nei ",
"Ling ",
"Qi ",
"Yue ",
"Lip ",
"Yi ",
"Xi ",
"Chen ",
"",
"Rong ",
"Chen ",
"Nong ",
"You ",
"Ji ",
"Bo ",
"Fang ",
"Gei ",
"",
"Cu ",
"Di ",
"Gaau ",
"Yu ",
"Ge ",
"Xu ",
"Lu ",
"He ",
"",
"Bai ",
"Gong ",
"Jiong ",
"Gwai ",
"Ya ",
"Nu ",
"You ",
"Song ",
"Xie ",
"Cang ",
"Yao ",
"Shu ",
"Yan ",
"Shuai ",
"Liao ",
"Sung ",
"Yu ",
"Bo ",
"Sui ",
"Cin ",
"Yan ",
"Lei ",
"Lin ",
"Tai ",
"Du ",
"Yue ",
"Ji ",
"Cin ",
"Yun ",
"Bong ",
"",
"",
"Ju ",
"Ceot ",
"Chen ",
"Cung ",
"Xiang ",
"Xian ",
"On ",
"Gui ",
"Yu ",
"Lei ",
"",
"Tu ",
"Chen ",
"Xing ",
"Qiu ",
"Hang ",
"",
"Dang ",
"Cai ",
"Di ",
"Yan ",
"Ci ",
"Gung ",
"Zing ",
"Chan ",
"",
"Li ",
"Suo ",
"Ma ",
"Ma ",
"",
"Tang ",
"Pei ",
"Lou ",
"Sat ",
"Cuo ",
"Tu ",
"E ",
"Can ",
"Jie ",
"Ti ",
"Ji ",
"Dang ",
"Jiao ",
"Bi ",
"Lei ",
"Yi ",
"Chun ",
"Chun ",
"Po ",
"Li ",
"Zai ",
"Tai ",
"Po ",
"Tian ",
"Ju ",
"Xu ",
"Fan ",
"",
"Xu ",
"Er ",
"Huo ",
"Chua ",
"Ran ",
"Fa ",
"Gyun ",
"",
"Liang ",
"Ti ",
"Mi ",
"",
],
"x49": [
"",
"Cen ",
"Mei ",
"Yin ",
"Mian ",
"Tu ",
"Kui ",
"Sau ",
"Hei ",
"Mi ",
"Rong ",
"Guo ",
"Coeng ",
"Mi ",
"Ju ",
"Pi ",
"Jin ",
"Wang ",
"Ji ",
"Meng ",
"Jian ",
"Xue ",
"Bao ",
"Gan ",
"Chan ",
"Li ",
"Li ",
"Qiu ",
"Dun ",
"Ying ",
"Yun ",
"Chen ",
"Ji ",
"Ran ",
"",
"Lue ",
"",
"Gui ",
"Yue ",
"Hui ",
"Pi ",
"Cha ",
"Duo ",
"Chan ",
"So ",
"Kuan ",
"She ",
"Xing ",
"Weng ",
"Shi ",
"Chi ",
"Ye ",
"Han ",
"Fei ",
"Ye ",
"Yan ",
"Zuan ",
"Saau ",
"Yin ",
"Duo ",
"Xian ",
"Gwaan ",
"Tou ",
"Qie ",
"Chan ",
"Han ",
"Meng ",
"Yue ",
"Cu ",
"Qian ",
"Jin ",
"Shan ",
"Mu ",
"Zyun ",
"Coeng ",
"Baang ",
"Zheng ",
"Zhi ",
"Chun ",
"Yu ",
"Mou ",
"Wan ",
"Chou ",
"Kei ",
"Su ",
"Pie ",
"Tian ",
"Kuan ",
"Cu ",
"Sui ",
"Co ",
"Jie ",
"Jian ",
"Ao ",
"Jiao ",
"Ye ",
"Saam ",
"Ye ",
"Long ",
"Zao ",
"Bao ",
"Lian ",
"",
"Huan ",
"Lu ",
"Wei ",
"Xian ",
"Tie ",
"Bo ",
"Zheng ",
"Zhu ",
"Ba ",
"Meng ",
"Xie ",
"Aau ",
"Zaau ",
"Teon ",
"Xiao ",
"Li ",
"Zha ",
"Mi ",
"",
"Ye ",
"",
"",
"Put ",
"Xie ",
"",
"",
"Bong ",
"Shan ",
"Coeng ",
"",
"Shan ",
"Jue ",
"Ji ",
"Fang ",
"",
"Niao ",
"Ao ",
"Chu ",
"Wu ",
"Guan ",
"Xie ",
"Ting ",
"Xie ",
"Dang ",
"Zim ",
"Tan ",
"Ping ",
"Xia ",
"Xu ",
"Bi ",
"Si ",
"Huo ",
"Zheng ",
"Wu ",
"",
"Run ",
"Chuai ",
"Shi ",
"Huan ",
"Kuo ",
"Fu ",
"Chuai ",
"Xian ",
"Qin ",
"Qie ",
"Lan ",
"",
"Ya ",
"Zing ",
"Que ",
"",
"Chun ",
"Zhi ",
"Gau ",
"Kui ",
"Qian ",
"Hang ",
"Yi ",
"Ni ",
"Zheng ",
"Chuai ",
"",
"Shi ",
"Cap ",
"Ci ",
"Jue ",
"Xu ",
"Yun ",
"",
"",
"Chu ",
"Dao ",
"Dian ",
"Ge ",
"Ti ",
"Hong ",
"Ni ",
"",
"Li ",
"",
"Xian ",
"Seoi ",
"Xi ",
"Xuan ",
"",
"Gwik ",
"Taai ",
"Lai ",
"Zau ",
"Mu ",
"Cheng ",
"Jian ",
"Bi ",
"Qi ",
"Ling ",
"Hao ",
"Bang ",
"Tang ",
"Di ",
"Fu ",
"Xian ",
"Shuan ",
"Zung ",
"Keoi ",
"Zaan ",
"Pu ",
"Hui ",
"Wei ",
"Yi ",
"Ye ",
"",
"Che ",
"Hao ",
"Baai ",
"",
"Xian ",
"Chan ",
"Hun ",
"Gaau ",
"Han ",
"Ci ",
"Zi ",
"Qi ",
"Kui ",
"Rou ",
"Gu ",
"Zing ",
"Xiong ",
"Gap ",
"Hu ",
"Cui ",
"Syu ",
"Que ",
],
"x4a": [
"Di ",
"Che ",
"Cau ",
"",
"Yan ",
"Liao ",
"Bi ",
"Soeng ",
"Ban ",
"Zing ",
"",
"Nue ",
"Bao ",
"Ying ",
"Hong ",
"Ci ",
"Qia ",
"Ti ",
"Yu ",
"Lei ",
"Bao ",
"Wu ",
"Ji ",
"Fu ",
"Xian ",
"Cen ",
"Fat ",
"Se ",
"Baang ",
"Cing ",
"Yu ",
"Waa ",
"Ai ",
"Han ",
"Dan ",
"Ge ",
"Di ",
"Hu ",
"Pang ",
"Zaam ",
"Zaa ",
"Ling ",
"Mai ",
"Mai ",
"Lian ",
"Siu ",
"Xue ",
"Zhen ",
"Po ",
"Fu ",
"Nou ",
"Xi ",
"Dui ",
"Dan ",
"Yun ",
"Xian ",
"Yin ",
"Suk ",
"Dui ",
"Beng ",
"Hu ",
"Fei ",
"Fei ",
"Qian ",
"Bei ",
"Fei ",
"",
"Shi ",
"Tian ",
"Zhan ",
"Jian ",
"",
"Hui ",
"Fu ",
"Wan ",
"Mo ",
"Qiao ",
"Liao ",
"",
"Mie ",
"Ge ",
"Hong ",
"Yu ",
"Qi ",
"Duo ",
"Ang ",
"Saa ",
"Ba ",
"Di ",
"Xuan ",
"Di ",
"Bi ",
"Zhou ",
"Pao ",
"Nian ",
"Yi ",
"Ting ",
"Jia ",
"Da ",
"Duo ",
"Xi ",
"Dan ",
"Tiao ",
"Xie ",
"Chang ",
"Yuan ",
"Guan ",
"Liang ",
"Beng ",
"Gei ",
"Lu ",
"Ji ",
"Xuan ",
"Shu ",
"Dou ",
"Shu ",
"Hu ",
"Yun ",
"Chan ",
"Bong ",
"Rong ",
"E ",
"Zung ",
"Ba ",
"Feng ",
"Zyu ",
"Zhe ",
"Fen ",
"Guan ",
"Bu ",
"Ge ",
"Deon ",
"Huang ",
"Du ",
"Ti ",
"Bo ",
"Qian ",
"La ",
"Long ",
"Wei ",
"Zhan ",
"Lan ",
"Seoi ",
"Na ",
"Bi ",
"Tuo ",
"Jiao ",
"",
"Bu ",
"Ju ",
"Po ",
"Xia ",
"Wei ",
"Fu ",
"He ",
"Fan ",
"Chan ",
"Hu ",
"Za ",
"",
"Saai ",
"Zai ",
"Zai ",
"Zai ",
"Fan ",
"Die ",
"Hong ",
"Chi ",
"Bao ",
"Yin ",
"",
"Haang ",
"Bo ",
"Ruan ",
"Chou ",
"Ying ",
"Zi ",
"Gai ",
"Kwaan ",
"Yun ",
"Zhen ",
"Ya ",
"Zeoi ",
"Hou ",
"Min ",
"Pei ",
"Ge ",
"Bian ",
"Zyut ",
"Hao ",
"Mi ",
"Sheng ",
"Gen ",
"Bi ",
"Duo ",
"Chun ",
"Chua ",
"San ",
"Cheng ",
"Ran ",
"Zen ",
"Mao ",
"Bo ",
"Tui ",
"Pi ",
"Fu ",
"Zyut ",
"Hei ",
"Lin ",
"Hei ",
"Men ",
"Wu ",
"Qi ",
"Zhi ",
"Chen ",
"Xia ",
"He ",
"Sang ",
"Gwaa ",
"Hou ",
"Au ",
"Fu ",
"Rao ",
"Hun ",
"Pei ",
"Qian ",
"Si ",
"Xi ",
"Ming ",
"Kui ",
"Ge ",
"",
"Ao ",
"San ",
"Shuang ",
"Lou ",
"Zhen ",
"Hui ",
"Can ",
"Ci ",
"Lin ",
"Na ",
"Han ",
"Du ",
"Jin ",
"Mian ",
"Fan ",
"E ",
"Nao ",
"Hong ",
"Hong ",
"Xue ",
"Xue ",
"Pau ",
"Bi ",
"Ciu ",
],
"x4b": [
"You ",
"Yi ",
"Xue ",
"Sa ",
"Yu ",
"Li ",
"Li ",
"Yuan ",
"Dui ",
"Hao ",
"Qie ",
"Leng ",
"Paau ",
"Fat ",
"Guo ",
"Bu ",
"Wei ",
"Wei ",
"Saau ",
"An ",
"Xu ",
"Shang ",
"Heng ",
"Yang ",
"Gwik ",
"Yao ",
"Fan ",
"Bi ",
"Ci ",
"Heng ",
"Tao ",
"Liu ",
"Fei ",
"Zhu ",
"Tou ",
"Qi ",
"Chao ",
"Yi ",
"Dou ",
"Yuan ",
"Cu ",
"Zai ",
"Bo ",
"Can ",
"Yang ",
"Tou ",
"Yi ",
"Nian ",
"Shao ",
"Ben ",
"Ngaau ",
"Ban ",
"Mo ",
"Ai ",
"En ",
"She ",
"Caan ",
"Zhi ",
"Yang ",
"Jian ",
"Yuan ",
"Dui ",
"Ti ",
"Wei ",
"Xun ",
"Zhi ",
"Yi ",
"Ren ",
"Shi ",
"Hu ",
"Ne ",
"Yi ",
"Jian ",
"Sui ",
"Ying ",
"Bao ",
"Hu ",
"Hu ",
"Xie ",
"",
"Yang ",
"Lian ",
"Sik ",
"En ",
"Deoi ",
"Jian ",
"Zhu ",
"Ying ",
"Yan ",
"Jin ",
"Chuang ",
"Dan ",
"",
"Kuai ",
"Yi ",
"Ye ",
"Jian ",
"En ",
"Ning ",
"Ci ",
"Qian ",
"Xue ",
"Bo ",
"Mi ",
"Shui ",
"Mi ",
"Liang ",
"Qi ",
"Qi ",
"Shou ",
"Bi ",
"Bo ",
"Beng ",
"Bie ",
"Ni ",
"Wei ",
"Huan ",
"Fan ",
"Qi ",
"Liu ",
"Fu ",
"Ang ",
"Ang ",
"Fan ",
"Qi ",
"Qun ",
"Tuo ",
"Yi ",
"Bo ",
"Pian ",
"Bo ",
"Keoi ",
"Xuan ",
"",
"Baai6",
"Yu ",
"Chi ",
"Lu ",
"Yi ",
"Li ",
"Zaau ",
"Niao ",
"Xi ",
"Wu ",
"Gwing ",
"Lei ",
"Pei ",
"Zhao ",
"Zui ",
"Chuo ",
"Coeng ",
"An ",
"Er ",
"Yu ",
"Leng ",
"Fu ",
"Sha ",
"Huan ",
"Chu ",
"Sou ",
"Bik ",
"Bi ",
"Die ",
"Song ",
"Di ",
"Li ",
"Giu ",
"Han ",
"Zai ",
"Gu ",
"Cheng ",
"Lou ",
"Mo ",
"Mi ",
"Mai ",
"Ao ",
"Dan ",
"Zhu ",
"Huang ",
"Fan ",
"Deng ",
"Tong ",
"",
"Du ",
"Hu ",
"Wei ",
"Ji ",
"Chi ",
"Lin ",
"Biu ",
"Pang ",
"Jian ",
"Nie ",
"Luo ",
"Ji ",
"Ngon ",
"Waa ",
"Nie ",
"Yi ",
"Gwaat6",
"Wan ",
"Ya ",
"Qia ",
"Bo ",
"Hau ",
"Ling ",
"Gan ",
"Huo ",
"Hai ",
"Hong ",
"Heng ",
"Kui ",
"Cen ",
"Ting ",
"Lang ",
"Bi ",
"Huan ",
"Po ",
"Ou ",
"Jian ",
"Ti ",
"Sui ",
"Kwaa ",
"Dui ",
"Ao ",
"Jian ",
"Mo ",
"Gui ",
"Kuai ",
"An ",
"Ma ",
"Qing ",
"Fen ",
"",
"Kao ",
"Hao ",
"Duo ",
"Cim ",
"Nai ",
"Seoi ",
"Jie ",
"Fu ",
"Pa ",
"Sung ",
"Chang ",
"Nie ",
"Man ",
"Sung ",
"Ci ",
"Cim ",
"Kuo ",
"Gai ",
"Di ",
"Fu ",
"Tiao ",
"Zu ",
],
"x4c": [
"Wo ",
"Fei ",
"Cai ",
"Peng ",
"Shi ",
"Sou ",
"Rou ",
"Qi ",
"Cha ",
"Pan ",
"Bo ",
"Man ",
"Zong ",
"Ci ",
"Gui ",
"Ji ",
"Lan ",
"Siu ",
"Meng ",
"Mian ",
"Pan ",
"Lu ",
"Cuan ",
"Gau ",
"Liu ",
"Yi ",
"Wen ",
"Li ",
"Li ",
"Zeng ",
"Zhu ",
"Hun ",
"Shen ",
"Chi ",
"Xing ",
"Wang ",
"Dung ",
"Huo ",
"Pi ",
"Bou6",
"Mei ",
"Che ",
"Mei ",
"Chao ",
"Ju ",
"Nou ",
"",
"Ni ",
"Ru ",
"Ling ",
"Ya ",
"",
"Qi ",
"Zi ",
"",
"Bang ",
"Gung ",
"Ze ",
"Jie ",
"Yu ",
"Xin ",
"Bei ",
"Ba ",
"Tuo ",
"Ong ",
"Qiao ",
"You ",
"Di ",
"Jie ",
"Mo ",
"Sheng ",
"Shan ",
"Qi ",
"Shan ",
"Mi ",
"Dan ",
"Yi ",
"Geng ",
"Geng ",
"Tou ",
"Fu ",
"Xue ",
"Yi ",
"Ting ",
"Tiao ",
"Mou ",
"Liu ",
"Caan ",
"Li ",
"Suk ",
"Lu ",
"Xu ",
"Cuo ",
"Ba ",
"Liu ",
"Ju ",
"Zhan ",
"Ju ",
"Zang ",
"Zu ",
"Xian ",
"Zhi ",
"",
"",
"Zhi ",
"",
"",
"La ",
"Seoi ",
"Geng ",
"E ",
"Mu ",
"Zhong ",
"Di ",
"Yan ",
"Zin ",
"Geng ",
"Zung ",
"Lang ",
"Yu ",
"Caau ",
"Na ",
"Hai ",
"Hua ",
"Zhan ",
"Coeng ",
"Lou ",
"Chan ",
"Die ",
"Wei ",
"Xuan ",
"Zao ",
"Min ",
"Kwai ",
"Sou ",
"",
"",
"Si ",
"Tuo ",
"Cen ",
"Kuan ",
"Teng ",
"Nei ",
"Lao ",
"Lu ",
"Yi ",
"Xie ",
"Yan ",
"Qing ",
"Pu ",
"Chou ",
"Xian ",
"Guan ",
"Jie ",
"Lai ",
"Meng ",
"Ye ",
"Ceoi ",
"Li ",
"Yin ",
"Ceon ",
"Cau ",
"Teng ",
"Yu ",
"",
"Gau ",
"Cha ",
"Du ",
"Hong ",
"Si ",
"Xi ",
"Gaau ",
"Qi ",
"Ci ",
"Yuan ",
"Ji ",
"Yun ",
"Fang ",
"Gung ",
"Hang ",
"Zhen ",
"Hu ",
"",
"",
"Jie ",
"Pei ",
"Gan ",
"Xuan ",
"Saang ",
"Dao ",
"Qiao ",
"Ci ",
"Die ",
"Ba ",
"Tiao ",
"Wan ",
"Ci ",
"Zhi ",
"Bai ",
"Wu ",
"Bao ",
"Dan ",
"Ba ",
"Tong ",
"Gyun ",
"Gung ",
"Jiu ",
"Gui ",
"Ci ",
"You ",
"Yuan ",
"Lao ",
"Jiu ",
"Fou ",
"Nei ",
"E ",
"E ",
"Xing ",
"He ",
"Yan ",
"Tu ",
"Bu ",
"Beng ",
"Kou ",
"Chui ",
"Zeoi ",
"Qi ",
"Yuan ",
"Bit ",
"",
"Hyun ",
"Hou ",
"Huang ",
"Ziu ",
"Juan ",
"Kui ",
"E ",
"Ji ",
"Mo ",
"Chong ",
"Bao ",
"Wu ",
"Zhen ",
"Xu ",
"Da ",
"Chi ",
"Gaai ",
"Cong ",
"Ma ",
"Kou ",
"Yan ",
"Can ",
"Aau ",
"He ",
"Dang ",
"Lan ",
],
"x4d": [
"Tong ",
"Yu ",
"Hang ",
"Nao ",
"Li ",
"Fen ",
"Pu ",
"Ling ",
"Ao ",
"Xuan ",
"Yi ",
"Xuan ",
"Meng ",
"Ang ",
"Lei ",
"Yan ",
"Bao ",
"Die ",
"Ling ",
"Shi ",
"Jiao ",
"Lie ",
"Jing ",
"Ju ",
"Ti ",
"Pi ",
"Gang ",
"Jiao ",
"Huai ",
"Bu ",
"Di ",
"Huan ",
"Yao ",
"Li ",
"Mi ",
"Fu ",
"Saang ",
"Gaa ",
"Ren ",
"Wai ",
"",
"Piao ",
"Lu ",
"Ling ",
"Yi ",
"Cai ",
"Shan ",
"Fat ",
"Shu ",
"Tuo ",
"Mo ",
"He ",
"Tie ",
"Bing ",
"Peng ",
"Hun ",
"Fu ",
"Guo ",
"Bu ",
"Li ",
"Chan ",
"Bai ",
"Cuo ",
"Meng ",
"Suo ",
"Qiang ",
"Zhi ",
"Kuang ",
"Bi ",
"Ao ",
"Meng ",
"Xian ",
"Guk ",
"Tou ",
"Teon ",
"Wei ",
"Cim ",
"Tan ",
"Caau ",
"Lao ",
"Chan ",
"Ni ",
"Ni ",
"Li ",
"Dong ",
"Ju ",
"Jian ",
"Fu ",
"Sha ",
"Zha ",
"Tao ",
"Jian ",
"Nong ",
"Ya ",
"Jing ",
"Gan ",
"Di ",
"Jian ",
"Mei ",
"Da ",
"Jian ",
"She ",
"Xie ",
"Zai ",
"Mang ",
"Li ",
"Gun ",
"Yu ",
"Ta ",
"Zhe ",
"Yang ",
"Tuan ",
"Soeng ",
"He ",
"Diao ",
"Wei ",
"Yun ",
"Zha ",
"Qu ",
"Waa ",
"Caau ",
"Zi ",
"Ting ",
"Gu ",
"Soeng ",
"Ca ",
"Fu ",
"Tie ",
"Ta ",
"Ta ",
"Zhuo ",
"Han ",
"Ping ",
"He ",
"Ceoi ",
"Zhou ",
"Bo ",
"Liu ",
"Nu ",
"Kaap ",
"Pao ",
"Di ",
"Sha ",
"Ti ",
"Kuai ",
"Ti ",
"Qi ",
"Ji ",
"Chi ",
"Pa ",
"Jin ",
"Ke ",
"Li ",
"Ju ",
"Qu ",
"La ",
"Gu ",
"Qia ",
"Qi ",
"Xian ",
"Jian ",
"Shi ",
"Xian ",
"Ai ",
"Hua ",
"Ju ",
"Ze ",
"Yao ",
"Tam ",
"Ji ",
"Cha ",
"Kan ",
"Gin ",
"",
"Yan ",
"Gwaai ",
"Ziu ",
"Tong ",
"Nan ",
"Yue ",
"Ceoi ",
"Chi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x4e": [
"Il ",
"Ceng ",
"Kyo ",
"Chil ",
"Sang ",
"Ha ",
"",
"Man ",
"Cang ",
"Sam ",
"Sang ",
"Ha ",
"Ki ",
"Pwu ",
"Ye ",
"Myen ",
"Kal ",
"Chwuk ",
"Chou ",
"Zhuan ",
"Cha ",
"Pi ",
"Sey ",
"Shi ",
"Kwu ",
"Pyeng ",
"Ye ",
"Cong ",
"Dong ",
"Si ",
"Sung ",
"Cwu ",
"Kwu ",
"Lyang ",
"Cwu ",
"Yu ",
"Lyang ",
"Yan ",
"Pyeng ",
"Sang ",
"Kon ",
"Kwu ",
"Kay ",
"A ",
"Qiang ",
"Cwung ",
"Kuk ",
"Jie ",
"Pong ",
"Kwan ",
"Koc ",
"Chan ",
"Lin ",
"Chak ",
"Cwu ",
"Ha ",
"Hwan ",
"Tan ",
"Wei ",
"Cwu ",
"Ceng ",
"Li ",
"Ju ",
"Pyel ",
"Pwul ",
"Yi ",
"Yey ",
"Nay ",
"Shime ",
"Kwu ",
"Kwu ",
"Cek ",
"Yo ",
"Yi ",
"",
"Ci ",
"Wu ",
"Sa ",
"Ho ",
"Phip ",
"Le ",
"Im ",
"Ping ",
"Pang ",
"Qiao ",
"Hu ",
"Koy ",
"Sung ",
"Sung ",
"Ul ",
"Yin ",
"",
"Mya ",
"Kwu ",
"Kel ",
"Ya ",
"Xi ",
"Xiang ",
"Gai ",
"Kyu ",
"Hal ",
"",
"Shu ",
"Twul ",
"Si ",
"Kyey ",
"Nang ",
"Kal ",
"Kel ",
"Tol ",
"",
"Ol ",
"Mai ",
"Lan ",
"Cal ",
"Yu ",
"Xue ",
"Yan ",
"Phol ",
"Sal ",
"Na ",
"Gan ",
"Sol ",
"El ",
"Cwul ",
"",
"Kan ",
"Chi ",
"Kwi ",
"Ken ",
"Lan ",
"Lin ",
"Uy ",
"Kwel ",
"Lyo ",
"Ma ",
"Ye ",
"Zheng ",
"Sa ",
"Sa ",
"I ",
"Chok ",
"Wu ",
"Yu ",
"Wul ",
"Wun ",
"Ho ",
"Ki ",
"O ",
"Ceng ",
"Si ",
"Sui ",
"Sen ",
"Kung ",
"Ya ",
"Sa ",
"A ",
"Cey ",
"A ",
"Kuk ",
"Twu ",
"Mang ",
"Hang ",
"Ta ",
"Kyo ",
"Hay ",
"Yek ",
"Chan ",
"Hyeng ",
"Mu ",
"",
"Hyang ",
"Kyeng ",
"Ceng ",
"Lyang ",
"Xiang ",
"Kyeng ",
"Ya ",
"Qin ",
"Pak ",
"Yu ",
"Xie ",
"Tan ",
"Lyem ",
"Duo ",
"Mi ",
"In ",
"In ",
"Ji ",
"La ",
"Mang ",
"Yi ",
"Sip ",
"In ",
"Luk ",
"Ding ",
"Chuk ",
"Jin ",
"Pwu ",
"Kwu ",
"Ba ",
"Zhang ",
"Kum ",
"Kay ",
"Bing ",
"Ing ",
"Cong ",
"Pwul ",
"San ",
"Lun ",
"Sya ",
"Cang ",
"Ca ",
"Sa ",
"Tha ",
"Cang ",
"Pwu ",
"Sen ",
"Hen ",
"Cha ",
"Hong ",
"Tong ",
"In ",
"Chen ",
"Gan ",
"Hul ",
"Cak ",
"Tay ",
"Lyeng ",
"I ",
"Chao ",
"Chang ",
"Sa ",
"",
"Yi ",
"Mu ",
"Men ",
"In ",
"Ka ",
"Chao ",
"Ang ",
"Qian ",
"Cwung ",
"Pi ",
"Wan ",
"O ",
"Ken ",
"Kay ",
"Yao ",
"Feng ",
"Cang ",
"Im ",
"Wang ",
"Pin ",
"Di ",
"Pang ",
],
"x4f": [
"Cong ",
"Ki ",
"Pei ",
"Ye ",
"Diao ",
"Ton ",
"Wen ",
"Yek ",
"Sim ",
"Hang ",
"I ",
"Kup ",
"Ai ",
"O ",
"Ki ",
"Pok ",
"Pel ",
"Hyu ",
"Jin ",
"Bei ",
"Dan ",
"Fu ",
"Tang ",
"Zhong ",
"You ",
"Hwa ",
"Hoy ",
"Yu ",
"Col ",
"Cen ",
"San ",
"Wei ",
"Chuan ",
"Che ",
"Ya ",
"Hyen ",
"Shang ",
"Chang ",
"Lun ",
"Cang ",
"Xun ",
"Sin ",
"Wei ",
"Zhu ",
"",
"Xuan ",
"Nu ",
"Payk ",
"Ko ",
"Ni ",
"Ni ",
"Xie ",
"Pan ",
"Xu ",
"Lyeng ",
"Zhou ",
"Sin ",
"Qu ",
"Sa ",
"Phayng ",
"Sa ",
"Ka ",
"Pi ",
"Yi ",
"Sa ",
"I ",
"Ceng ",
"Cen ",
"Han ",
"Mai ",
"Tan ",
"Ce ",
"Pho ",
"Kwu ",
"Phi ",
"So ",
"Ci ",
"Wi ",
"Ce ",
"Cwu ",
"Cwa ",
"Wu ",
"Ang ",
"Pwun ",
"Cem ",
"Ha ",
"Phil ",
"Tha ",
"Sa ",
"Ye ",
"Il ",
"Pwul ",
"Cak ",
"Kwu ",
"Nyeng ",
"Tong ",
"Ni ",
"Xuan ",
"Qu ",
"Yong ",
"Wa ",
"Qian ",
"",
"Ka ",
"",
"Phay ",
"Hoy ",
"He ",
"Lao ",
"Xiang ",
"Ge ",
"Yang ",
"Peyk ",
"Fa ",
"Ming ",
"Ka ",
"I ",
"Pyeng ",
"Kil ",
"Hang ",
"Kwal ",
"Kwey ",
"Cen ",
"Co ",
"Kyo ",
"Cha ",
"Il ",
"Sa ",
"Hyeng ",
"Sin ",
"Thak ",
"Kan ",
"Cil ",
"Hay ",
"Lay ",
"I ",
"Chi ",
"Kwa ",
"Kwang ",
"Lyey ",
"Yin ",
"Si ",
"Mi ",
"Cwu ",
"Hyek ",
"Yu ",
"An ",
"Lu ",
"Mo ",
"Er ",
"Lyun ",
"Tong ",
"Cha ",
"Chik ",
"Swun ",
"Kong ",
"Cwu ",
"Uy ",
"Ye ",
"Jian ",
"Hyep ",
"Ka ",
"Zai ",
"Lye ",
"Ko ",
"Jiao ",
"Zhen ",
"Ce ",
"Qiao ",
"Kuai ",
"Chai ",
"Ning ",
"Nong ",
"Jin ",
"Mo ",
"Hwu ",
"Jiong ",
"Cheng ",
"Cin ",
"Cwa ",
"Chou ",
"Chim ",
"Lye ",
"Kwuk ",
"Twu ",
"Ceng ",
"Shen ",
"Thal ",
"Bo ",
"Nan ",
"Hao ",
"Phyen ",
"Tui ",
"Yu ",
"Kyey ",
"Chok ",
"A ",
"Kwu ",
"Xu ",
"Kwang ",
"Ku ",
"O ",
"Cwun ",
"Up ",
"Po ",
"Lyang ",
"Co ",
"So ",
"Li ",
"Yong ",
"Hun ",
"Kyeng ",
"Hyen ",
"San ",
"Pai ",
"Sok ",
"Pwu ",
"Huy ",
"Li ",
"Myen ",
"Ping ",
"Po ",
"Yu ",
"Sa ",
"Hyep ",
"Sin ",
"Xiu ",
"O ",
"Cey ",
"Ke ",
"Chou ",
"",
"Yan ",
"Lia ",
"Li ",
"Lai ",
"Si ",
"Jian ",
"Swu ",
"Pwu ",
"Hwa ",
"Kwu ",
"Hyo ",
"Pay ",
"Chen ",
"Phyo ",
"Swuk ",
"Pi ",
"Pong ",
"A ",
"Em ",
"Pi ",
"Yu ",
"Xin ",
"Pi ",
"Jian ",
],
"x50": [
"Chang ",
"Chi ",
"Pyeng ",
"Kwu ",
"Hyo ",
"Swi ",
"Lyang ",
"Wan ",
"Lai ",
"Chang ",
"Zong ",
"Kay ",
"Kwan ",
"Pay ",
"Cen ",
"Swuk ",
"Swuk ",
"Mwun ",
"To ",
"Tam ",
"Kwul ",
"Swu ",
"Hayng ",
"Pwung ",
"Tang ",
"Hwu ",
"Uy ",
"Ki ",
"Chek ",
"Kan ",
"Kyeng ",
"Cha ",
"Hyu ",
"Chang ",
"Chep ",
"Pang ",
"Chi ",
"Kong ",
"Kwen ",
"Cong ",
"Ke ",
"Chen ",
"Yey ",
"Lyun ",
"Thak ",
"Wi ",
"La ",
"Song ",
"Lung ",
"Hon ",
"Tong ",
"Sa ",
"Pwun ",
"Wu ",
"Kwu ",
"Nai ",
"Cai ",
"Kem ",
"Zhai ",
"Ya ",
"Chi ",
"Sha ",
"Qing ",
"",
"Ying ",
"Ching ",
"Jian ",
"En ",
"Nan ",
"Tong ",
"Chwun ",
"Ka ",
"Key ",
"Wi ",
"Yu ",
"Bing ",
"Ya ",
"Ti ",
"Oy ",
"Phyen ",
"An ",
"Feng ",
"Thang ",
"Ak ",
"E ",
"Hay ",
"Cha ",
"Sheng ",
"Kan ",
"Di ",
"Cwu ",
"Sap ",
"Ceng ",
"Pay ",
"Yep ",
"Hwang ",
"Yo ",
"Zhan ",
"Cho ",
"Yan ",
"You ",
"Ken ",
"Xu ",
"Zha ",
"Chi ",
"Pwu ",
"Phip ",
"Chi ",
"Chong ",
"Myen ",
"Ji ",
"Uy ",
"Sel ",
"Xun ",
"Si ",
"Duan ",
"Chuk ",
"Ceng ",
"Wu ",
"Thwu ",
"Thwu ",
"Pi ",
"Za ",
"Lu ",
"Jie ",
"Wi ",
"Fen ",
"Chang ",
"Koy ",
"Sou ",
"Zhi ",
"So ",
"Xia ",
"Pwu ",
"Yuan ",
"Rong ",
"Li ",
"Ru ",
"Yun ",
"Kang ",
"Ma ",
"Pang ",
"Cen ",
"Tang ",
"Hao ",
"Kel ",
"Hyey ",
"Sen ",
"Kyem ",
"Kak ",
"Chang ",
"Chu ",
"San ",
"Pi ",
"Hyo ",
"Yong ",
"Yo ",
"Thap ",
"Sa ",
"Yang ",
"Fa ",
"Bing ",
"Ka ",
"Dai ",
"Zai ",
"Tang ",
"Cot ",
"Bin ",
"Chu ",
"Nuo ",
"Can ",
"Lei ",
"Choy ",
"Yong ",
"Co ",
"Chong ",
"Pwung ",
"Song ",
"O ",
"Cen ",
"Kwu ",
"Chay ",
"Cou ",
"Sang ",
"Qiang ",
"Jing ",
"Cey ",
"Sa ",
"Han ",
"Cang ",
"Kyeng ",
"En ",
"Di ",
"Sel ",
"Lwu ",
"Bei ",
"Phyo ",
"Kun ",
"Lyen ",
"Lyuk ",
"Man ",
"Chem ",
"Sen ",
"Tham ",
"Ying ",
"Tong ",
"Cwun ",
"Sang ",
"Shan ",
"Kyo ",
"Kwun ",
"Tui ",
"Cwun ",
"Pok ",
"Huy ",
"Lao ",
"Chang ",
"Guang ",
"Lyo ",
"Ki ",
"Deng ",
"Can ",
"Wi ",
"Ki ",
"Fan ",
"Hui ",
"Chen ",
"Cham ",
"Than ",
"Yo ",
"Chwu ",
"Sung ",
"Pwun ",
"Han ",
"Kyel ",
"E ",
"Cho ",
"Cham ",
"Tong ",
"Lin ",
"Pwuk ",
"Gu ",
"",
"Su ",
"Xian ",
"Kang ",
"Min ",
"Yep ",
"Kum ",
"Ka ",
"Qiao ",
"Pyek ",
"Feng ",
"Chwu ",
"Ay ",
"Sa ",
],
"x51": [
"Uy ",
"Cwun ",
"Nong ",
"Chen ",
"Ek ",
"Dang ",
"Kyeng ",
"Hyen ",
"Koy ",
"Kem ",
"Chu ",
"Tam ",
"Kyo ",
"Sha ",
"Zai ",
"",
"Pin ",
"Ap ",
"Yu ",
"Tay ",
"Cwu ",
"Cey ",
"Lan ",
"Uy ",
"Cin ",
"Qian ",
"Meng ",
"Mwu ",
"Nyeng ",
"Qiong ",
"Ni ",
"Sang ",
"Lyep ",
"Loy ",
"Lye ",
"Kuang ",
"Pho ",
"Yuk ",
"Phyo ",
"Chan ",
"Cil ",
"Sa ",
"Wu ",
"Ho ",
"Chen ",
"Chun ",
"Li ",
"Teng ",
"Wei ",
"Long ",
"Ce ",
"Cham ",
"Sang ",
"Swuk ",
"Hyey ",
"Lye ",
"La ",
"Chan ",
"Na ",
"Tang ",
"Em ",
"Lay ",
"Nang ",
"In ",
"Ol ",
"Yun ",
"Cam ",
"Wen ",
"Hyeng ",
"Chwung ",
"Co ",
"Hyung ",
"Sen ",
"Kwang ",
"Thay ",
"Kuk ",
"Thay ",
"Myen ",
"Tho ",
"Chang ",
"A ",
"Thay ",
"A ",
"Chim ",
"Tho ",
"Si ",
"Yen ",
"Yen ",
"Shi ",
"Sip ",
"Tang ",
"Chen ",
"Twu ",
"Fen ",
"Mao ",
"Sin ",
"Dou ",
"Payk ",
"Kung ",
"Li ",
"Huang ",
"Ip ",
"Mang ",
"Nay ",
"Cen ",
"Lyang ",
"Yu ",
"Phal ",
"Kong ",
"Lyuk ",
"Hyey ",
"",
"Lan ",
"Kong ",
"Tian ",
"Guan ",
"Xing ",
"Pyeng ",
"Ki ",
"Kwu ",
"Cen ",
"Ca ",
"Ppwun ",
"Yang ",
"Kyem ",
"Shou ",
"Ji ",
"Yi ",
"Ki ",
"Chan ",
"Kyeng ",
"Mo ",
"Yem ",
"Nay ",
"Wen ",
"Mao ",
"Gang ",
"Yem ",
"Chayk ",
"Jiong ",
"Chayk ",
"Cay ",
"Gua ",
"Kyeng ",
"Mo ",
"Cwu ",
"Mo ",
"Kwu ",
"Hwu ",
"Myen ",
"Myek ",
"Yong ",
"Yu ",
"Sa ",
"Kan ",
"Jun ",
"Nong ",
"Uy ",
"Mi ",
"Sek ",
"Kwan ",
"Mong ",
"Chong ",
"Chwi ",
"Wen ",
"Myeng ",
"Kou ",
"Lam ",
"Pwu ",
"Sa ",
"Myek ",
"Ping ",
"Tong ",
"Tai ",
"Gang ",
"Feng ",
"Ping ",
"Ho ",
"Chwung ",
"Kyel ",
"Ho ",
"Hwang ",
"Ya ",
"Layng ",
"Pan ",
"Pwul ",
"Min ",
"Dong ",
"Xian ",
"Lyel ",
"Xia ",
"Jian ",
"Jing ",
"Swu ",
"Mei ",
"Tu ",
"Che ",
"Ko ",
"Cwun ",
"Song ",
"Ceng ",
"Lyang ",
"Cheng ",
"Co ",
"Lung ",
"Tong ",
"Gan ",
"Kam ",
"Yin ",
"Cwu ",
"Uy ",
"Yul ",
"Chang ",
"Ming ",
"Cwun ",
"Choy ",
"Si ",
"Duo ",
"Kum ",
"Lum ",
"Lum ",
"Ung ",
"Huy ",
"Du ",
"Kwey ",
"Pem ",
"Pem ",
"Pem ",
"Feng ",
"Ke ",
"Che ",
"Tako ",
"Phwung ",
"Mok ",
"Ci ",
"Fu ",
"Feng ",
"Ping ",
"Phwung ",
"Kai ",
"Hwang ",
"Kay ",
"Gan ",
"Tung ",
"Ping ",
"Kam ",
"Hyung ",
"Koy ",
"Chel ",
"Yo ",
"Chwul ",
"Ji ",
"Dang ",
"Ham ",
"Ham ",
"Zao ",
],
"x52": [
"To ",
"Co ",
"To ",
"In ",
"In ",
"Chang ",
"Pwun ",
"Cel ",
"Yey ",
"Ji ",
"Kan ",
"Chen ",
"Chon ",
"Chu ",
"Mwun ",
"Ki ",
"Dan ",
"Hyeng ",
"Hwa ",
"Wan ",
"Kyel ",
"Li ",
"Wel ",
"Lyel ",
"Lyu ",
"Ze ",
"Gang ",
"Chuang ",
"Pwul ",
"Cho ",
"Qu ",
"Ju ",
"San ",
"Min ",
"Ling ",
"Zhong ",
"Phan ",
"Pyel ",
"Kep ",
"Kep ",
"Pho ",
"Li ",
"San ",
"Pyel ",
"Chan ",
"Jing ",
"Kwal ",
"Gen ",
"To ",
"Chang ",
"Kyu ",
"Ko ",
"Tha ",
"I ",
"Cey ",
"Sway ",
"Kwen ",
"Chal ",
"Ca ",
"Kak ",
"Jie ",
"Gui ",
"Ci ",
"Gui ",
"Kai ",
"Duo ",
"Ji ",
"Chey ",
"Kyeng ",
"Lou ",
"Gen ",
"Chik ",
"Yuan ",
"Cwa ",
"Sak ",
"Kuk ",
"Lal ",
"Cen ",
"Chal ",
"Chang ",
"Gua ",
"Jian ",
"Chak ",
"Li ",
"Chek ",
"Pi ",
"Pwu ",
"Can ",
"Qi ",
"Chang ",
"Sa ",
"Kang ",
"Wan ",
"Pak ",
"Ki ",
"Chel ",
"Kyeng ",
"Sem ",
"Zhuo ",
"Kem ",
"Ji ",
"Pak ",
"Em ",
"Ju ",
"Koyk ",
"Ing ",
"Cen ",
"Thak ",
"Tan ",
"Ok ",
"Kwa ",
"Pwu ",
"Sheng ",
"Jian ",
"Hal ",
"Cha ",
"Kay ",
"Chang ",
"Juan ",
"San ",
"Tan ",
"Lyuk ",
"Li ",
"Fou ",
"Shan ",
"Phyo ",
"Kou ",
"Cho ",
"Kwal ",
"Qiao ",
"Kwey ",
"Hoyk ",
"Cha ",
"Zhuo ",
"Lian ",
"Kuk ",
"Pyek ",
"Lyu ",
"Hoy ",
"Co ",
"Kwi ",
"Kem ",
"Kem ",
"Tang ",
"Hwak ",
"Cey ",
"Kem ",
"Uy ",
"Kem ",
"Zhi ",
"Cham ",
"Chan ",
"Ma ",
"Li ",
"Chok ",
"Lyek ",
"Ya ",
"Quan ",
"Ban ",
"Kong ",
"Ka ",
"Wu ",
"Mai ",
"Lyel ",
"Kun ",
"Keng ",
"Hyep ",
"Zhi ",
"Dong ",
"Co ",
"No ",
"Kep ",
"Kwu ",
"Cho ",
"Il ",
"Zhu ",
"Miao ",
"Lye ",
"Jing ",
"Lao ",
"Lo ",
"Kwen ",
"Kwu ",
"Yang ",
"Wa ",
"Hyo ",
"Mou ",
"Kwang ",
"Hal ",
"Lie ",
"Hayk ",
"Shi ",
"Kuk ",
"Kyeng ",
"Hao ",
"Pal ",
"Min ",
"Chik ",
"Lang ",
"Yong ",
"Yong ",
"Myen ",
"Ke ",
"Xun ",
"Kwen ",
"Kyeng ",
"Lu ",
"Pou ",
"Meng ",
"Lay ",
"Luk ",
"Kai ",
"Min ",
"Tong ",
"Wuk ",
"Wuk ",
"Kam ",
"Mwu ",
"Yi ",
"Hwun ",
"Weng ",
"Sung ",
"Lo ",
"Mo ",
"Lyuk ",
"Phyo ",
"Sey ",
"Cek ",
"Kun ",
"Qiang ",
"Cho ",
"Kwen ",
"Yang ",
"Yey ",
"Jue ",
"Fan ",
"Juan ",
"Tong ",
"Ke ",
"Tan ",
"Hyep ",
"May ",
"Hwun ",
"Hwun ",
"Lye ",
"Lye ",
"Chel ",
"Yang ",
"Kwen ",
"Pho ",
"Cak ",
"Kyun ",
"Kwu ",
"Mwun ",
"Kwu ",
"Mwul ",
],
"x53": [
"Kyun ",
"Mwun ",
"Nay ",
"Kay ",
"Kay ",
"Pho ",
"Chong ",
"",
"Hung ",
"Phyeng ",
"Kwuk ",
"To ",
"Hap ",
"Pho ",
"Ap ",
"Pho ",
"Pok ",
"Kwung ",
"Tap ",
"Kwu ",
"Qiong ",
"Pi ",
"Hwa ",
"Pwuk ",
"Noy ",
"Si ",
"Pang ",
"Kwu ",
"I ",
"Cap ",
"Cang ",
"Kang ",
"Cang ",
"Kwang ",
"Hu ",
"Kap ",
"Qu ",
"Pyen ",
"Gui ",
"Hyep ",
"Zang ",
"Kuang ",
"Pi ",
"Hu ",
"Yu ",
"Kwey ",
"Gui ",
"Hoy ",
"Tan ",
"Kwey ",
"Lyem ",
"Lyem ",
"San ",
"Tok ",
"Kwu ",
"Qu ",
"Hyey ",
"Phil ",
"Kwu ",
"Yey ",
"Ap ",
"En ",
"Phyen ",
"Nik ",
"Kwu ",
"Sip ",
"Sin ",
"Chen ",
"Ip ",
"Sap ",
"Col ",
"Sung ",
"O ",
"Hwey ",
"Pan ",
"Sey ",
"Sip ",
"Man ",
"Hua ",
"Xie ",
"Man ",
"Pi ",
"Col ",
"Thak ",
"Hyep ",
"Tan ",
"Mai ",
"Nam ",
"Tan ",
"Cip ",
"Pak ",
"Shuai ",
"Pok ",
"Kwan ",
"Pyen ",
"Key ",
"Cem ",
"Cap ",
"Lu ",
"Yu ",
"Lu ",
"Xi ",
"Kway ",
"Wo ",
"Sel ",
"Cel ",
"Jie ",
"Wei ",
"Ang ",
"Kong ",
"Chi ",
"Myo ",
"In ",
"Wi ",
"So ",
"Cuk ",
"Kak ",
"Lan ",
"Si ",
"Kwen ",
"Sa ",
"Hyul ",
"Kun ",
"Kak ",
"Ol ",
"Cuk ",
"Ak ",
"Kyeng ",
"Sul ",
"",
"Em ",
"Cem ",
"Ayk ",
"Ting ",
"Li ",
"Chayk ",
"Han ",
"Lye ",
"A ",
"Ap ",
"Yan ",
"She ",
"Ci ",
"Zha ",
"Pang ",
"",
"He ",
"Ay ",
"Cil ",
"Ce ",
"Pang ",
"Ti ",
"Li ",
"She ",
"Hwu ",
"Ting ",
"Zui ",
"Co ",
"Pi ",
"Wen ",
"Chuk ",
"Yuan ",
"Xiang ",
"Yan ",
"Lyek ",
"Kwel ",
"Ha ",
"Dian ",
"Cwu ",
"Kwu ",
"Kun ",
"Ao ",
"Kwey ",
"Yem ",
"Si ",
"Li ",
"Chang ",
"Em ",
"Lye ",
"Em ",
"Yan ",
"Wen ",
"Sa ",
"Koyng ",
"Lin ",
"Kwu ",
"Ke ",
"Ke ",
"Uk ",
"Lwi ",
"Du ",
"Xian ",
"Zhuan ",
"Sam ",
"Sam ",
"Cham ",
"Cham ",
"Cham ",
"Ai ",
"Dai ",
"Wu ",
"Cha ",
"Kup ",
"Wu ",
"Ssang ",
"Pan ",
"Shou ",
"Kway ",
"Ba ",
"Fa ",
"Yak ",
"Sa ",
"Swuk ",
"Chel ",
"Chwi ",
"Swu ",
"Pyen ",
"Se ",
"Ka ",
"Pan ",
"Sou ",
"Gao ",
"Wi ",
"Swu ",
"Die ",
"Yey ",
"Chong ",
"Kwu ",
"Ko ",
"Kwu ",
"Lyeng ",
"Gua ",
"To ",
"Ko ",
"Ci ",
"Kyu ",
"So ",
"Phal ",
"Ceng ",
"Ka ",
"Thay ",
"Cil ",
"Sa ",
"Wu ",
"Kwu ",
"Pha ",
"Hyep ",
"Ho ",
"Sa ",
"Tan ",
"Ip ",
"Le ",
"Diao ",
"Ji ",
"Duk ",
"Hong ",
],
"x54": [
"Mie ",
"Yu ",
"Mang ",
"Hul ",
"Kak ",
"Hyen ",
"Yo ",
"Zi ",
"Hap ",
"Kil ",
"Cek ",
"Chon ",
"Tong ",
"Myeng ",
"Hwu ",
"Li ",
"Tho ",
"Hyang ",
"Tha ",
"Xia ",
"Ye ",
"Lye ",
"A ",
"Ma ",
"Wu ",
"Xue ",
"Uy ",
"Kwun ",
"Chou ",
"Lin ",
"Than ",
"Um ",
"Phyey ",
"Phil ",
"Sim ",
"Qin ",
"Jie ",
"Bu ",
"Pwu ",
"Pha ",
"Twun ",
"Pwun ",
"Wa ",
"Ham ",
"Un ",
"Hang ",
"Cen ",
"Kyey ",
"Hong ",
"Ci ",
"Sin ",
"O ",
"O ",
"Myo ",
"Nwul ",
"Hyel ",
"Hup ",
"Chwi ",
"Twu ",
"Mwun ",
"Hwu ",
"Hwu ",
"O ",
"Ko ",
"Ha ",
"Jun ",
"Lye ",
"Ay ",
"Ge ",
"May ",
"May ",
"Qi ",
"Ceng ",
"O ",
"Ko ",
"Fu ",
"Kyu ",
"Hong ",
"Chek ",
"Sung ",
"Nwul ",
"Than ",
"Fu ",
"Yi ",
"Dai ",
"Ou ",
"Li ",
"Bai ",
"Yuan ",
"Kuai ",
"",
"Qiang ",
"Wu ",
"Ayk ",
"Shi ",
"Hyen ",
"Pen ",
"Wen ",
"Ni ",
"Bou ",
"Lyeng ",
"Ran ",
"Yu ",
"Cey ",
"Cwu ",
"Shi ",
"Cwu ",
"Chep ",
"Hi ",
"Yey ",
"Ki ",
"Ping ",
"Ca ",
"Ko ",
"Zi ",
"Mi ",
"Kwu ",
"Ka ",
"No ",
"Hap ",
"Pi ",
"Il ",
"Hyo ",
"Sin ",
"Ho ",
"Myeng ",
"Tal ",
"Ke ",
"Ce ",
"Kan ",
"Cap ",
"Tuo ",
"Tol ",
"Pou ",
"Pho ",
"Phil ",
"Pwul ",
"Ang ",
"Hwa ",
"Sa ",
"Hwa ",
"Hay ",
"Kwu ",
"Yeng ",
"Pwu ",
"Que ",
"Cwu ",
"Wa ",
"Ka ",
"Ko ",
"Ka ",
"Zuo ",
"Bu ",
"Long ",
"Dong ",
"Ning ",
"Tha ",
"Si ",
"Xian ",
"Huo ",
"Cip ",
"I ",
"Ak ",
"Guang ",
"Tha ",
"Huy ",
"I ",
"Lyel ",
"Ca ",
"Mi ",
"Mi ",
"Ci ",
"Kyo ",
"Kil ",
"Cwu ",
"Kak ",
"Shuai ",
"Chal ",
"So ",
"Hay ",
"Hui ",
"Kua ",
"Si ",
"To ",
"Ham ",
"E ",
"Hwen ",
"Hyu ",
"Kway ",
"In ",
"Lao ",
"I ",
"Ay ",
"Phwum ",
"Sin ",
"Tong ",
"Hong ",
"Hyung ",
"Chi ",
"Wa ",
"Hap ",
"Cay ",
"Yu ",
"Di ",
"Pai ",
"Sang ",
"Ay ",
"Hen ",
"Kuang ",
"Ya ",
"Da ",
"Xiao ",
"Bi ",
"Yue ",
"",
"Hua ",
"Sasou ",
"Kuai ",
"Duo ",
"",
"Ji ",
"Nong ",
"Mou ",
"Yo ",
"Hao ",
"Wen ",
"Long ",
"Pou ",
"Pang ",
"Ka ",
"A ",
"Chi ",
"Cho ",
"Li ",
"Na ",
"Cok ",
"Ha ",
"Kok ",
"Hyo ",
"Hyen ",
"Lo ",
"Pal ",
"Chel ",
"Chal ",
"Liang ",
"Ba ",
"Ma ",
"Lyel ",
"Sui ",
"Pwu ",
"Pho ",
"Han ",
"Hyeng ",
"Kyeng ",
"Shuo ",
"Ka ",
],
"x55": [
"Yu ",
"En ",
"Gu ",
"Gu ",
"Phay ",
"Ham ",
"Sa ",
"Cin ",
"Up ",
"Ay ",
"Kyep ",
"Tu ",
"Yen ",
"Wan ",
"Li ",
"Huy ",
"Tang ",
"Zuo ",
"Qiu ",
"Che ",
"O ",
"Co ",
"A ",
"Dou ",
"Qi ",
"Di ",
"Chim ",
"Ma ",
"Mal ",
"Hong ",
"Dou ",
"Kes ",
"Lao ",
"Liang ",
"Suo ",
"Zao ",
"Hwan ",
"Laang ",
"Sha ",
"Ji ",
"Zuo ",
"Wo ",
"Pong ",
"Kum ",
"Ho ",
"Ki ",
"Swu ",
"Yu ",
"Shua ",
"Chang ",
"A ",
"Lye ",
"Kang ",
"An ",
"Cha ",
"Yuk ",
"Cem ",
"E ",
"Tian ",
"Lai ",
"Sap ",
"Xi ",
"Tha ",
"Hol ",
"Ay ",
"Co ",
"Nyo ",
"Sap ",
"Thak ",
"Thak ",
"Sang ",
"Cek ",
"Heng ",
"Lam ",
"A ",
"Kyo ",
"Xiang ",
"Thon ",
"O ",
"Mwun ",
"Chway ",
"Cap ",
"Hu ",
"Kyey ",
"Qi ",
"To ",
"Tam ",
"Tam ",
"Ye ",
"Ca ",
"Pi ",
"Cui ",
"Chel ",
"Wa ",
"A ",
"Kyey ",
"Zhe ",
"Pay ",
"Liang ",
"Ham ",
"Pi ",
"Sha ",
"La ",
"Ze ",
"Qing ",
"Gua ",
"Pa ",
"Zhe ",
"Se ",
"Zhuan ",
"Nie ",
"Guo ",
"Luo ",
"Yan ",
"Di ",
"Quan ",
"Tan ",
"Bo ",
"Ding ",
"Lang ",
"Xiao ",
"Geoi ",
"Tang ",
"Si ",
"Cey ",
"Am ",
"Chwu ",
"Tam ",
"Kayk ",
"Ong ",
"Wei ",
"Nam ",
"Sen ",
"Yuk ",
"Chel ",
"Lal ",
"Kay ",
"Hwu ",
"Ham ",
"Chep ",
"Cwu ",
"Cay ",
"Wai ",
"Ya ",
"Yu ",
"Um ",
"Cam ",
"Yo ",
"Ak ",
"Mian ",
"Hu ",
"Wun ",
"Chen ",
"Hwey ",
"Hwan ",
"Hwan ",
"Huy ",
"Kal ",
"Cul ",
"Wi ",
"Cong ",
"Wei ",
"Sha ",
"Hwu ",
"Hwang ",
"Du ",
"Yep ",
"Hwen ",
"Lyang ",
"Yu ",
"Sang ",
"Kkik ",
"Kyo ",
"An ",
"Tan ",
"Pen ",
"Sik ",
"Li ",
"Yak ",
"Sa ",
"Wei ",
"Miao ",
"Yeng ",
"Pen ",
"Phos ",
"Kui ",
"Xi ",
"Yu ",
"Jie ",
"Lou ",
"Ku ",
"So ",
"Hak ",
"Cey ",
"Yo ",
"Hyo ",
"Sa ",
"Hwu ",
"Chang ",
"Sayk ",
"Ong ",
"So ",
"Hong ",
"Hyep ",
"I ",
"Suo ",
"Ma ",
"Cha ",
"Hal ",
"Hap ",
"Thap ",
"Sang ",
"Cin ",
"Yok ",
"Sou ",
"Wa ",
"Ji ",
"Pang ",
"O ",
"Kyem ",
"Ki ",
"Kyek ",
"Ca ",
"Cha ",
"La ",
"Ong ",
"Ol ",
"Sa ",
"Chi ",
"Hao ",
"Suo ",
"Lyun ",
"Hai ",
"Sway ",
"Qin ",
"Nie ",
"He ",
"Cis ",
"Sai ",
"Ng ",
"Ge ",
"Na ",
"Dia ",
"Ai ",
"",
"Tong ",
"Phil ",
"O ",
"O ",
"Lyen ",
"Choy ",
"Cha ",
"Mak ",
"Swu ",
"Cwu ",
"Tham ",
],
"x56": [
"Di ",
"Chuk ",
"Kyo ",
"Chong ",
"Kyo ",
"Kay ",
"Than ",
"Sam ",
"Co ",
"Ka ",
"Ay ",
"Xiao ",
"Phyo ",
"Lwu ",
"Al ",
"Ha ",
"Kyo ",
"Ho ",
"Hyey ",
"Koyk ",
"Kwu ",
"En ",
"Chayk ",
"Sang ",
"He ",
"Po ",
"De ",
"Ma ",
"Ma ",
"Hu ",
"Luk ",
"To ",
"Al ",
"Tang ",
"Ye ",
"Beng ",
"Ying ",
"Saai ",
"Jiao ",
"Mi ",
"So ",
"Hwa ",
"Mai ",
"Yen ",
"Choy ",
"Peng ",
"Cho ",
"So ",
"Ki ",
"Chok ",
"Co ",
"Koy ",
"Chwi ",
"Hyo ",
"Si ",
"Ho ",
"Mwu ",
"Lyo ",
"Kyo ",
"Huy ",
"Hyu ",
"Than ",
"Tam ",
"Mwuk ",
"Son ",
"O ",
"Cwun ",
"Pen ",
"Chi ",
"Wi ",
"Cham ",
"Tam ",
"Chwuk ",
"Tam ",
"Yul ",
"Thon ",
"Cayng ",
"Cho ",
"Yel ",
"Hup ",
"Ki ",
"Hao ",
"Lian ",
"He ",
"Deng ",
"Wi ",
"Yin ",
"Pu ",
"Jue ",
"Qin ",
"Xun ",
"Nie ",
"Lu ",
"Si ",
"Em ",
"Ying ",
"Tal ",
"Dan ",
"Wuk ",
"Cwu ",
"Kum ",
"Nong ",
"El ",
"Hway ",
"Ki ",
"Ak ",
"Co ",
"Huy ",
"Se ",
"Kyo ",
"Yen ",
"Ay ",
"Ong ",
"Kyak ",
"Khway ",
"Wu ",
"Pwun ",
"Dao ",
"Kal ",
"Xin ",
"Thon ",
"Dang ",
"Sin ",
"Sai ",
"Pi ",
"Pi ",
"Yin ",
"Zui ",
"Nyeng ",
"Di ",
"Lam ",
"Thap ",
"Hoyk ",
"Yu ",
"Hyo ",
"Ha ",
"Ya ",
"Duo ",
"Pi ",
"Chou ",
"Cey ",
"Cin ",
"Ho ",
"Chey ",
"Chang ",
"",
"",
"Ca ",
"Chey ",
"Lo ",
"Hui ",
"Pak ",
"Wu ",
"Kyo ",
"Un ",
"Hak ",
"Mwuk ",
"Huang ",
"Zhe ",
"Li ",
"Lyu ",
"Haai ",
"Nang ",
"Xiao ",
"Mo ",
"Yen ",
"Lyek ",
"Lo ",
"Long ",
"Fu ",
"Tam ",
"Chin ",
"Pin ",
"Pi ",
"Hyang ",
"Huo ",
"Mo ",
"Huy ",
"Tha ",
"Ko ",
"Em ",
"Chan ",
"Ayng ",
"Yang ",
"Dian ",
"La ",
"Ta ",
"Hyo ",
"Cak ",
"Chuo ",
"Hwan ",
"Huo ",
"Cen ",
"Sep ",
"Hyo ",
"Cap ",
"Li ",
"Chen ",
"Chay ",
"Li ",
"Yey ",
"La ",
"Nang ",
"Chal ",
"Su ",
"Huy ",
"So ",
"Kan ",
"Cal ",
"Chok ",
"Lan ",
"Sel ",
"Nang ",
"Trom ",
"Lo ",
"Kwuk ",
"Hui ",
"In ",
"Swu ",
"Sa ",
"Nin ",
"Ken ",
"Hoy ",
"Sin ",
"In ",
"Nan ",
"Tuan ",
"Tan ",
"Ton ",
"Kang ",
"Yen ",
"Kyeng ",
"Pian ",
"Wun ",
"Chang ",
"Hol ",
"Hoy ",
"Wan ",
"Wa ",
"Guo ",
"Kon ",
"Chong ",
"Wi ",
"To ",
"Wei ",
"Lun ",
"Kwuk ",
"Kyun ",
"Ri ",
"Lyeng ",
"Ko ",
"Guo ",
"Tai ",
"Kwuk ",
"Tu ",
"Yu ",
],
"x57": [
"Kwuk ",
"Un ",
"Hon ",
"Pho ",
"E ",
"Ham ",
"Yuan ",
"Lyun ",
"Kwen ",
"E ",
"Cheng ",
"Kwuk ",
"Chen ",
"Wi ",
"Wen ",
"Kwen ",
"Ku ",
"Fu ",
"Wen ",
"Wen ",
"E ",
"Toshokan ",
"To ",
"To ",
"Tan ",
"Lue ",
"Hui ",
"Yek ",
"Wen ",
"Lan ",
"Luan ",
"Tho ",
"Ya ",
"Tu ",
"Ting ",
"Sheng ",
"Pu ",
"Lok ",
"Iri ",
"Ap ",
"Cay ",
"Wei ",
"Ge ",
"Yu ",
"O ",
"Kyu ",
"Pi ",
"Yi ",
"Ci ",
"Qian ",
"Qian ",
"Zhen ",
"Zhuo ",
"Dang ",
"Qia ",
"Akutsu ",
"Yama ",
"Kuang ",
"Chang ",
"Ki ",
"Nie ",
"Mo ",
"Kup ",
"Jia ",
"Ci ",
"Zhi ",
"Phan ",
"Xun ",
"Tou ",
"Qin ",
"Fen ",
"Kyun ",
"Keng ",
"Tun ",
"Pang ",
"Fen ",
"Pwun ",
"Tam ",
"Kam ",
"Pay ",
"Cwa ",
"Kayng ",
"Bi ",
"Xing ",
"Di ",
"Jing ",
"Ji ",
"Kuai ",
"Di ",
"Jing ",
"Jian ",
"Tan ",
"Li ",
"Ba ",
"Wu ",
"Fen ",
"Zhui ",
"Pha ",
"Pan ",
"Tang ",
"Kon ",
"Qu ",
"Than ",
"Chek ",
"I ",
"Kam ",
"Phyeng ",
"Cem ",
"Gua ",
"Ni ",
"Tay ",
"Pay ",
"Kyeng ",
"Ang ",
"Fo ",
"Yo ",
"Liu ",
"Kwu ",
"Mu ",
"Ka ",
"Kwu ",
"Xue ",
"Ba ",
"Ci ",
"Thak ",
"Ling ",
"Zhu ",
"Fu ",
"Hu ",
"Zhi ",
"Swu ",
"La ",
"Long ",
"Long ",
"Lu ",
"Ao ",
"Tay ",
"Pao ",
"",
"Hyeng ",
"Tong ",
"Ji ",
"Ke ",
"Lu ",
"Ci ",
"Chi ",
"Lei ",
"Hay ",
"Yin ",
"Hwu ",
"Dui ",
"Zhao ",
"Fu ",
"Guang ",
"Yao ",
"Duo ",
"Duo ",
"Gui ",
"Tha ",
"Yang ",
"Un ",
"Fa ",
"Kwu ",
"Wen ",
"Cil ",
"Xie ",
"Ken ",
"Jiong ",
"Shou ",
"Seng ",
"Ha ",
"Dian ",
"Hong ",
"Wu ",
"Kua ",
"",
"Tao ",
"Dang ",
"Kai ",
"Gake ",
"Nao ",
"An ",
"Xing ",
"Xian ",
"Wan ",
"Bang ",
"Pei ",
"Ba ",
"Yi ",
"Un ",
"Han ",
"Xu ",
"Chui ",
"Cen ",
"Geng ",
"Ay ",
"Peng ",
"Fang ",
"Kak ",
"Yong ",
"Cwun ",
"Hyep ",
"Di ",
"May ",
"Lang ",
"Xuan ",
"Seng ",
"Yen ",
"Jin ",
"Chel ",
"Lal ",
"Lie ",
"Pu ",
"Cheng ",
"Gomi ",
"Bu ",
"Shi ",
"Xun ",
"Guo ",
"Jiong ",
"Ya ",
"Nian ",
"Di ",
"Yek ",
"Pwu ",
"Ya ",
"Juan ",
"Sui ",
"Pi ",
"Cheng ",
"Wan ",
"Ju ",
"Lun ",
"Zheng ",
"Kong ",
"Chong ",
"Dong ",
"Thay ",
"Tan ",
"An ",
"Chay ",
"Shu ",
"Beng ",
"Kam ",
"Sik ",
"Duo ",
"Yi ",
"Cip ",
"Yi ",
"Pay ",
"Ki ",
"Cwun ",
"Ki ",
"So ",
"Ju ",
"Ni ",
],
"x58": [
"Kwul ",
"Ke ",
"Tang ",
"Kon ",
"Ni ",
"Kyen ",
"Thoy ",
"Kun ",
"Kang ",
"Yuk ",
"Ak ",
"Pwul ",
"Gu ",
"Tu ",
"Leng ",
"",
"Ya ",
"Qian ",
"",
"An ",
"",
"Tha ",
"Nao ",
"Tol ",
"Cheng ",
"In ",
"Hun ",
"Bi ",
"Lian ",
"Kwa ",
"Chep ",
"Zhuan ",
"Hwu ",
"Po ",
"Bao ",
"Yu ",
"Cey ",
"Mao ",
"Kyey ",
"Yen ",
"E ",
"Geng ",
"Kam ",
"Zong ",
"Yu ",
"Huang ",
"E ",
"Yo ",
"En ",
"Po ",
"Ji ",
"Mei ",
"Cang ",
"To ",
"Tuo ",
"Yin ",
"Feng ",
"Zhong ",
"Kyey ",
"Zhen ",
"Feng ",
"Gang ",
"Chuan ",
"Jian ",
"Pyeng ",
"Rui ",
"Xiang ",
"Huang ",
"Leng ",
"Duan ",
"",
"Xuan ",
"Ji ",
"Chek ",
"Koy ",
"Yeng ",
"Thap ",
"Sup ",
"Yong ",
"Kay ",
"So ",
"So ",
"Si ",
"Mi ",
"Thap ",
"Weng ",
"Cheng ",
"To ",
"Tang ",
"Ko ",
"Chong ",
"Li ",
"Pong ",
"Bang ",
"Say ",
"Zang ",
"Dui ",
"Cen ",
"O ",
"Cheng ",
"Hwun ",
"Ge ",
"Zhen ",
"Ai ",
"Gong ",
"Yem ",
"Kan ",
"Cen ",
"Yuan ",
"Wen ",
"Xie ",
"Liu ",
"Ama ",
"Lang ",
"Chang ",
"Peng ",
"Beng ",
"Cin ",
"Cu ",
"Lu ",
"Ou ",
"Cham ",
"Mei ",
"Mo ",
"Cen ",
"Sang ",
"Swuk ",
"Lou ",
"Ci ",
"Man ",
"Biao ",
"Kyeng ",
"Qi ",
"Se ",
"Di ",
"Zhang ",
"Kan ",
"Yong ",
"Cem ",
"Chen ",
"Zhi ",
"Ki ",
"Guo ",
"Qiang ",
"Kun ",
"Di ",
"Shang ",
"Myo ",
"Cui ",
"Yan ",
"Ta ",
"Cung ",
"Qi ",
"Qiang ",
"Liang ",
"",
"Chwu ",
"Yo ",
"Cung ",
"He ",
"Shan ",
"Sen ",
"Ba ",
"Pok ",
"Kuai ",
"Dong ",
"Fan ",
"Que ",
"Mwuk ",
"Ton ",
"Dun ",
"Cwun ",
"Di ",
"Sheng ",
"Tha ",
"Duo ",
"Tam ",
"Tung ",
"Wu ",
"Pwun ",
"Huang ",
"Tan ",
"Da ",
"Ye ",
"Sho ",
"Mama ",
"O ",
"Cang ",
"Ji ",
"Qiao ",
"Kan ",
"Yi ",
"Pi ",
"Pyek ",
"Dian ",
"Kang ",
"Ye ",
"Ong ",
"Bo ",
"Tan ",
"Lan ",
"Ju ",
"Huai ",
"Dang ",
"Yang ",
"Qian ",
"Hwun ",
"Lan ",
"Xi ",
"Hak ",
"Ai ",
"Ap ",
"To ",
"Ho ",
"Ruan ",
"Mama ",
"Lwu ",
"Kwang ",
"Lo ",
"Yan ",
"Tam ",
"Yu ",
"Koy ",
"Long ",
"Long ",
"Rui ",
"Li ",
"Lin ",
"Yang ",
"Ten ",
"Xun ",
"Yan ",
"Lei ",
"Pha ",
"",
"Sa ",
"Im ",
"",
"Cang ",
"Cang ",
"Seng ",
"Il ",
"May ",
"Ke ",
"Zhu ",
"Zhuang ",
"Hu ",
"Hu ",
"Kun ",
"Il ",
"Ho ",
"Se ",
"Ho ",
"Swu ",
"Mang ",
"Zun ",
],
"x59": [
"Shou ",
"Yi ",
"Zhi ",
"Gu ",
"Chu ",
"Jiang ",
"Pong ",
"Bei ",
"Cay ",
"Pyen ",
"Sui ",
"Qun ",
"Ling ",
"Fu ",
"Zuo ",
"Ha ",
"Hyeng ",
"",
"Nao ",
"Xia ",
"Ki ",
"Sek ",
"Oy ",
"Yuan ",
"Mao ",
"Swuk ",
"Ta ",
"Duo ",
"Ya ",
"Qing ",
"Uys ",
"Gou ",
"Gou ",
"Qi ",
"Mong ",
"Meng ",
"In ",
"Kwa ",
"Chen ",
"Tay ",
"Ze ",
"Chen ",
"Thay ",
"Pwu ",
"Khway ",
"Yo ",
"Ang ",
"Hang ",
"Gao ",
"Sil ",
"Ben ",
"Tai ",
"Tou ",
"Yan ",
"Bi ",
"I ",
"Kwa ",
"Jia ",
"Duo ",
"Kwu ",
"Kuang ",
"Yun ",
"Hyep ",
"Pa ",
"En ",
"Lian ",
"Huan ",
"Di ",
"Em ",
"Pao ",
"Quan ",
"Ki ",
"Nay ",
"Pong ",
"Xie ",
"Fen ",
"Cem ",
"",
"Kyu ",
"Cwu ",
"Hwan ",
"Kyey ",
"Kai ",
"Cha ",
"Pwun ",
"Hyek ",
"Jiang ",
"Thwu ",
"Cang ",
"Ben ",
"Hay ",
"Xiang ",
"Fei ",
"Diao ",
"Xun ",
"Keng ",
"Cen ",
"Ao ",
"Sa ",
"Weng ",
"Pan ",
"O ",
"Wu ",
"O ",
"Cang ",
"Lyem ",
"Thal ",
"Yun ",
"Cang ",
"Sek ",
"Pwun ",
"Huo ",
"Bi ",
"Lian ",
"Duo ",
"Nye ",
"No ",
"Ding ",
"Nay ",
"Qian ",
"Kan ",
"Ta ",
"Jiu ",
"Nan ",
"Cha ",
"Ho ",
"Xian ",
"Fan ",
"Ji ",
"Cak ",
"Ye ",
"Pi ",
"Mang ",
"Hong ",
"Zhuang ",
"Fu ",
"Ma ",
"Dan ",
"Im ",
"Fu ",
"Jing ",
"Yen ",
"Xie ",
"Wen ",
"Zhong ",
"Pha ",
"Thwu ",
"Ki ",
"Keng ",
"Zhong ",
"Yo ",
"Kum ",
"Yun ",
"Myo ",
"Pei ",
"Shi ",
"Yue ",
"Cang ",
"Niu ",
"Yan ",
"Na ",
"Xin ",
"Fen ",
"Pi ",
"Yu ",
"Tha ",
"Feng ",
"Wan ",
"Pang ",
"Wu ",
"Yu ",
"Gui ",
"Thwu ",
"Ba ",
"Ni ",
"Chwuk ",
"Zhuo ",
"Zhao ",
"Tal ",
"Nai ",
"Yuan ",
"Tou ",
"Xuan ",
"Zhi ",
"E ",
"May ",
"Mo ",
"Che ",
"Bi ",
"Shen ",
"Chep ",
"E ",
"He ",
"Xu ",
"Fa ",
"Ceng ",
"Min ",
"Ban ",
"Mo ",
"Fu ",
"Lyeng ",
"Ca ",
"Zi ",
"Si ",
"Ran ",
"San ",
"Yang ",
"Man ",
"Ce ",
"Ko ",
"Sa ",
"Seng ",
"Wi ",
"Zi ",
"Ju ",
"San ",
"Pin ",
"Im ",
"Yo ",
"Tong ",
"Kang ",
"Cwu ",
"Kil ",
"Gai ",
"Shang ",
"Kuo ",
"Yen ",
"Kyo ",
"Kwu ",
"Mo ",
"Kan ",
"Jian ",
"I ",
"Nian ",
"Cil ",
"Huy ",
"Huy ",
"Xian ",
"Hang ",
"Guang ",
"Jun ",
"Kua ",
"Yan ",
"Ming ",
"Lie ",
"Pei ",
"Yan ",
"You ",
"Yen ",
"Cha ",
"Sin ",
"In ",
"Chi ",
"Gui ",
"Quan ",
"Ca ",
],
"x5a": [
"Song ",
"Wi ",
"Hong ",
"Way ",
"Lou ",
"Ya ",
"Rao ",
"Jiao ",
"Luan ",
"Ping ",
"Xian ",
"Shao ",
"Li ",
"Cheng ",
"Xiao ",
"Mang ",
"",
"Sa ",
"Wu ",
"Mi ",
"Ke ",
"Lai ",
"Chuo ",
"Ding ",
"Nang ",
"Hyeng ",
"Nan ",
"O ",
"Na ",
"Pei ",
"Nei ",
"Yen ",
"Sin ",
"Zhi ",
"Han ",
"Cey ",
"Zhuang ",
"A ",
"Pin ",
"Thay ",
"Han ",
"Man ",
"Wu ",
"Yan ",
"Wu ",
"Ay ",
"Yan ",
"O ",
"Si ",
"Yu ",
"Wa ",
"",
"Xian ",
"Chwu ",
"Chwi ",
"Shui ",
"Qi ",
"Xian ",
"Zhui ",
"Dong ",
"Chang ",
"Lu ",
"Ai ",
"A ",
"A ",
"Lwu ",
"Mian ",
"Cong ",
"Pou ",
"Ju ",
"Pha ",
"Cai ",
"Ding ",
"Wan ",
"Biao ",
"Xiao ",
"Swuk ",
"Qi ",
"Hui ",
"Fu ",
"Wa ",
"Wo ",
"Tan ",
"Fei ",
"",
"Jie ",
"Tian ",
"Ni ",
"Quan ",
"Jing ",
"Hon ",
"Jing ",
"Qian ",
"Dian ",
"Xing ",
"Hu ",
"Wa ",
"Lai ",
"Pi ",
"In ",
"Zhou ",
"Chuo ",
"Pwu ",
"Ceng ",
"Lun ",
"Yan ",
"Lam ",
"Kun ",
"Um ",
"Ya ",
"",
"Li ",
"Dian ",
"Xian ",
"Hwa ",
"Hua ",
"Ying ",
"Chan ",
"Shen ",
"Ceng ",
"Dang ",
"Yao ",
"Wu ",
"Nan ",
"Ruo ",
"Jia ",
"Tou ",
"Se ",
"Yu ",
"Wei ",
"Ti ",
"Rou ",
"Mi ",
"Dan ",
"Ruan ",
"Qin ",
"",
"Wu ",
"Qian ",
"Chun ",
"Mao ",
"Pwu ",
"Jie ",
"Duan ",
"Xi ",
"Zhong ",
"May ",
"Hwang ",
"Mian ",
"Am ",
"Ying ",
"Xuan ",
"",
"Wei ",
"Mi ",
"Wen ",
"Zhen ",
"Qiu ",
"Ti ",
"Sel ",
"Tuo ",
"Lian ",
"Mo ",
"Ran ",
"Si ",
"Pian ",
"Wei ",
"Way ",
"Jiu ",
"Hu ",
"O ",
"",
"Ho ",
"Xu ",
"Tou ",
"Gui ",
"Zou ",
"Yao ",
"Pi ",
"Sik ",
"Yuan ",
"Ing ",
"Rong ",
"Ru ",
"Chi ",
"Liu ",
"Mi ",
"Pan ",
"On ",
"Ma ",
"Kwu ",
"Koy ",
"Qin ",
"Ka ",
"Swu ",
"Zhen ",
"Wen ",
"Cha ",
"Yong ",
"Ming ",
"Ayng ",
"Cil ",
"Su ",
"Nyo ",
"Hyem ",
"Tao ",
"Pang ",
"Lang ",
"Nao ",
"Bao ",
"Ai ",
"Pi ",
"Pin ",
"Yi ",
"Phyo ",
"Kwu ",
"Lei ",
"Sen ",
"Man ",
"Yey ",
"Zhang ",
"Kang ",
"Yong ",
"Ni ",
"Li ",
"Cek ",
"Kyu ",
"En ",
"Jin ",
"Cen ",
"Hang ",
"Ce ",
"Han ",
"Nwun ",
"Lao ",
"Mo ",
"Zhe ",
"Ho ",
"Ho ",
"O ",
"Nen ",
"Qiang ",
"Maa ",
"Pie ",
"Gu ",
"Wu ",
"Cho ",
"Tuo ",
"Zhan ",
"Mao ",
"Han ",
"Han ",
"Mo ",
"Lyo ",
"Lian ",
"Hua ",
],
"x5b": [
"Kyu ",
"Deng ",
"Zhi ",
"Xu ",
"",
"Hwa ",
"Xi ",
"Hui ",
"Yo ",
"Huy ",
"Yan ",
"Sen ",
"Kyo ",
"Mei ",
"Fan ",
"Fan ",
"Xian ",
"Yi ",
"Wei ",
"Jiao ",
"Fu ",
"Shi ",
"Phyey ",
"Shan ",
"Sui ",
"Cang ",
"Lian ",
"Hyen ",
"",
"Niao ",
"Dong ",
"Yi ",
"Can ",
"Ai ",
"Yang ",
"Neng ",
"Ma ",
"Co ",
"Chou ",
"Jin ",
"Ca ",
"Yu ",
"Pin ",
"",
"Xu ",
"Nay ",
"Yan ",
"Tai ",
"Yeng ",
"Can ",
"Nyo ",
"",
"Yeng ",
"Mian ",
"Kaka ",
"Ma ",
"Shen ",
"Xing ",
"Ni ",
"Du ",
"Liu ",
"Yuan ",
"Lan ",
"Yen ",
"Sang ",
"Ling ",
"Jiao ",
"Yang ",
"Lan ",
"Sem ",
"Ying ",
"Shuang ",
"Shuai ",
"Quan ",
"Mi ",
"Li ",
"Lyen ",
"Yan ",
"Zhu ",
"Lan ",
"Ca ",
"Hyel ",
"Jue ",
"Jue ",
"Kong ",
"Ing ",
"Ca ",
"Ca ",
"Con ",
"Sun ",
"Pwu ",
"Phay ",
"Ca ",
"Hyo ",
"Xin ",
"Mayng ",
"Si ",
"Thay ",
"Bao ",
"Kyey ",
"Ko ",
"No ",
"Hak ",
"",
"Zhuan ",
"Hay ",
"Luan ",
"Son ",
"Huai ",
"Mie ",
"Cong ",
"Qian ",
"Swuk ",
"Can ",
"Ya ",
"Ca ",
"Ni ",
"Pwu ",
"Ca ",
"Li ",
"Hak ",
"Bo ",
"Yu ",
"Lai ",
"El ",
"El ",
"Ying ",
"San ",
"Mian ",
"Ce ",
"Yong ",
"Ta ",
"Gui ",
"Tayk ",
"Qiong ",
"Wu ",
"Swu ",
"An ",
"Ka ",
"Song ",
"Wan ",
"Rou ",
"Yao ",
"Koyng ",
"Yi ",
"Jing ",
"Zhun ",
"Pok ",
"Zhu ",
"Thang ",
"Hoyng ",
"Cong ",
"Kwan ",
"Cwu ",
"Ceng ",
"Wan ",
"Uy ",
"Po ",
"Shi ",
"Sil ",
"Chong ",
"Shen ",
"Kayk ",
"Sen ",
"Sil ",
"Yu ",
"Hwan ",
"Yi ",
"Tiao ",
"Shi ",
"Xian ",
"Kwung ",
"Seng ",
"Qun ",
"Kwung ",
"Xiao ",
"Cay ",
"Zha ",
"Bao ",
"Hay ",
"Yen ",
"So ",
"Ka ",
"Shen ",
"Sin ",
"Yong ",
"Huang ",
"Mil ",
"Kou ",
"Kuan ",
"Bin ",
"Swuk ",
"Chay ",
"Zan ",
"Cek ",
"Wen ",
"Ki ",
"In ",
"Mil ",
"Kwu ",
"Qing ",
"Que ",
"Zhen ",
"Jian ",
"Pwu ",
"Ning ",
"Bing ",
"Huan ",
"May ",
"Qin ",
"Han ",
"Wu ",
"Sik ",
"Nyeng ",
"Chim ",
"Nyeng ",
"Chi ",
"Yu ",
"Bao ",
"Kwan ",
"Nyeng ",
"Chim ",
"Mak ",
"Chal ",
"Ju ",
"Kwa ",
"Chim ",
"Hu ",
"O ",
"Lyo ",
"Sil ",
"Nyeng ",
"Chay ",
"Sim ",
"Wei ",
"Sa ",
"Kwan ",
"Hyey ",
"Lyo ",
"Cwun ",
"Hwan ",
"Yi ",
"Yi ",
"Po ",
"Qin ",
"Chong ",
"Po ",
"Feng ",
"Chon ",
"Tay ",
"Sa ",
"Xun ",
"Dao ",
"Lu ",
"Tay ",
"Swu ",
],
"x5c": [
"Po ",
"Pong ",
"Cen ",
"Fu ",
"Sa ",
"Kuk ",
"Cang ",
"Cang ",
"Cen ",
"Wi ",
"Con ",
"Sim ",
"Cwu ",
"Tay ",
"To ",
"So ",
"Ji ",
"So ",
"Er ",
"I ",
"I ",
"Ga ",
"Chem ",
"Shu ",
"Chen ",
"Sang ",
"Sang ",
"",
"Ga ",
"Chang ",
"Liao ",
"Sen ",
"Sen ",
"",
"Wang ",
"Wang ",
"Wu ",
"Liao ",
"Liao ",
"Yao ",
"Pang ",
"Wang ",
"Wang ",
"Wang ",
"Ga ",
"Yo ",
"Duo ",
"Kui ",
"Zhong ",
"Chwi ",
"Gan ",
"Gu ",
"Gan ",
"Tui ",
"Gan ",
"Gan ",
"Si ",
"Yun ",
"Chek ",
"Ko ",
"Ni ",
"Cin ",
"Mi ",
"Nyo ",
"Kwuk ",
"Pi ",
"Ceng ",
"Xi ",
"Bi ",
"Ke ",
"Kyey ",
"Cen ",
"Kwul ",
"Ti ",
"Jie ",
"Ok ",
"Diao ",
"Si ",
"Si ",
"Pyeng ",
"Kuk ",
"Sel ",
"Chen ",
"Xi ",
"Ni ",
"Cen ",
"Xi ",
"",
"Man ",
"E ",
"Lou ",
"Pyeng ",
"Ti ",
"Fei ",
"Chok ",
"Xie ",
"To ",
"Lu ",
"Lwu ",
"Xi ",
"Chung ",
"Li ",
"Ju ",
"Xie ",
"Kwu ",
"Jue ",
"Liao ",
"Jue ",
"Sok ",
"Xi ",
"Che ",
"Twun ",
"Ni ",
"San ",
"",
"Sen ",
"Li ",
"Xue ",
"Nata ",
"",
"Long ",
"Hul ",
"Ki ",
"Ren ",
"Wu ",
"Han ",
"Shen ",
"Yu ",
"Chwul ",
"Sui ",
"Qi ",
"",
"Yue ",
"Ban ",
"Yao ",
"Ang ",
"Ha ",
"Wu ",
"Cel ",
"E ",
"Kup ",
"Qian ",
"Fen ",
"Wan ",
"Ki ",
"Cam ",
"Kyem ",
"Qi ",
"Cha ",
"Jie ",
"Qu ",
"Gang ",
"Xian ",
"Ao ",
"Lan ",
"Dao ",
"Ba ",
"Cak ",
"Zuo ",
"Yang ",
"Ju ",
"Kang ",
"Ke ",
"Kwu ",
"Xue ",
"Pha ",
"Lip ",
"Cho ",
"Ce ",
"Am ",
"Pwul ",
"Swu ",
"Kap ",
"Lyeng ",
"Tuo ",
"Pei ",
"You ",
"Tay ",
"Kuang ",
"Ak ",
"Qu ",
"Ho ",
"Po ",
"Min ",
"An ",
"Tiao ",
"Lyeng ",
"Chi ",
"Yuri ",
"Dong ",
"Cem ",
"Kui ",
"Swu ",
"Mao ",
"Tong ",
"Xue ",
"Yi ",
"Kura ",
"He ",
"Ke ",
"Luo ",
"E ",
"Fu ",
"Swun ",
"Die ",
"Lu ",
"An ",
"Er ",
"Gai ",
"Quan ",
"Tong ",
"Yi ",
"Mu ",
"Shi ",
"An ",
"Wei ",
"Hu ",
"Chi ",
"Mi ",
"Li ",
"Ji ",
"Tong ",
"Wei ",
"You ",
"Sang ",
"Xia ",
"Li ",
"Yo ",
"Jiao ",
"Zheng ",
"Luan ",
"Jiao ",
"A ",
"A ",
"Yu ",
"Ye ",
"Bu ",
"Cho ",
"Qun ",
"Pong ",
"Pong ",
"No ",
"Li ",
"You ",
"Hyen ",
"Hong ",
"To ",
"Shen ",
"Cheng ",
"Tu ",
"Geng ",
"Cwun ",
"Hao ",
"Hyep ",
"Yin ",
"Yu ",
],
"x5d": [
"Lang ",
"Kan ",
"Lao ",
"Lai ",
"Hem ",
"Que ",
"Kong ",
"Swung ",
"Chong ",
"Ta ",
"",
"Hua ",
"Ju ",
"Lay ",
"Ki ",
"Min ",
"Kon ",
"Kon ",
"Zu ",
"Gu ",
"Choy ",
"Ay ",
"Ay ",
"Kang ",
"Lun ",
"Lyun ",
"Leng ",
"Kwul ",
"Duo ",
"Zheng ",
"Guo ",
"Um ",
"Dong ",
"Han ",
"Cayng ",
"Wei ",
"Hyo ",
"Pi ",
"Em ",
"Swung ",
"Jie ",
"Pwung ",
"Zu ",
"Jue ",
"Dong ",
"Zhan ",
"Gu ",
"Yin ",
"",
"Ze ",
"Huang ",
"Yu ",
"Oy ",
"Yang ",
"Feng ",
"Qiu ",
"Dun ",
"Ti ",
"Yi ",
"Zhi ",
"Shi ",
"Cay ",
"Yao ",
"E ",
"Zhu ",
"Kam ",
"Lyul ",
"Yan ",
"Mei ",
"Gan ",
"Ji ",
"Ji ",
"Huan ",
"Ting ",
"Sheng ",
"Mi ",
"Kam ",
"Wu ",
"Yu ",
"Zong ",
"Lam ",
"Jue ",
"Am ",
"Am ",
"Oy ",
"Zong ",
"Cha ",
"Sui ",
"Rong ",
"Yamashina ",
"Kum ",
"Yu ",
"Ki ",
"Lou ",
"Tu ",
"Dui ",
"Xi ",
"Weng ",
"Cang ",
"Dang ",
"Hong ",
"Jie ",
"Ai ",
"Liu ",
"Wu ",
"Swung ",
"Qiao ",
"Zi ",
"Oy ",
"Beng ",
"Dian ",
"Cha ",
"Qian ",
"Yong ",
"Nie ",
"Cuo ",
"Ji ",
"",
"Tao ",
"Song ",
"Zong ",
"Jiang ",
"Liao ",
"",
"Chan ",
"Die ",
"Cen ",
"Ding ",
"Tu ",
"Lwu ",
"Cang ",
"Zhan ",
"Cham ",
"Ao ",
"Cao ",
"Kwu ",
"Qiang ",
"Zui ",
"Zui ",
"To ",
"Dao ",
"Xi ",
"Yu ",
"Bo ",
"Long ",
"Xiang ",
"Cung ",
"Bo ",
"Kum ",
"Cho ",
"Yan ",
"Lao ",
"Zhan ",
"Lin ",
"Liao ",
"Liao ",
"Jin ",
"Tung ",
"Duo ",
"Zun ",
"Kyo ",
"Gui ",
"Yo ",
"Qiao ",
"Yao ",
"Jue ",
"Zhan ",
"Yek ",
"Xue ",
"Nao ",
"Ep ",
"Ep ",
"Yi ",
"E ",
"Hem ",
"Ji ",
"Hay ",
"Ke ",
"Xi ",
"Di ",
"Ao ",
"Zui ",
"",
"Uy ",
"Yeng ",
"Dao ",
"Lyeng ",
"Za ",
"Se ",
"Ak ",
"Yin ",
"",
"Jie ",
"Li ",
"Sui ",
"Long ",
"Long ",
"Dian ",
"Ying ",
"Xi ",
"Ju ",
"Cham ",
"Ying ",
"Kyu ",
"Yan ",
"Oy ",
"Nao ",
"Quan ",
"Chao ",
"Chan ",
"Man ",
"Cen ",
"Cen ",
"",
"Am ",
"Yan ",
"Yan ",
"Nao ",
"Hen ",
"Chuan ",
"Gui ",
"Chen ",
"Cwu ",
"Huang ",
"Jing ",
"Swun ",
"So ",
"Chao ",
"Lie ",
"Kong ",
"Cwa ",
"Kyo ",
"Ke ",
"Gong ",
"Kek ",
"Mwu ",
"Pwu ",
"Pwu ",
"Cha ",
"Qiu ",
"Qiu ",
"Ki ",
"I ",
"Sa ",
"Pha ",
"Chi ",
"Zhao ",
"Hang ",
"Yi ",
"Kun ",
"Son ",
"Kwen ",
"Phas ",
"Son ",
"Ken ",
"Fu ",
],
"x5e": [
"Za ",
"Bi ",
"Si ",
"Pho ",
"Ding ",
"Shuai ",
"Pem ",
"Nie ",
"Shi ",
"Pwun ",
"Pa ",
"Zhi ",
"Huy ",
"Hu ",
"Dan ",
"Wei ",
"Zhang ",
"Thang ",
"Dai ",
"Ma ",
"Pei ",
"Mal ",
"Chep ",
"Fu ",
"Lyem ",
"Cil ",
"Chwu ",
"Payk ",
"Zhi ",
"Cey ",
"Mo ",
"Yi ",
"Yi ",
"Ping ",
"Qia ",
"Juan ",
"Ru ",
"Sol ",
"Tay ",
"Zheng ",
"Sey ",
"Qiao ",
"Zhen ",
"Sa ",
"Qun ",
"Sek ",
"Bang ",
"Tay ",
"Kwi ",
"Chou ",
"Ping ",
"Cang ",
"Sha ",
"Wan ",
"Tay ",
"Yu ",
"Sang ",
"Sha ",
"Qi ",
"Ze ",
"Guo ",
"Mo ",
"Du ",
"Hwu ",
"Ceng ",
"Xu ",
"Mi ",
"Wi ",
"Ak ",
"Phok ",
"Yi ",
"Pang ",
"Ping ",
"Tazuna ",
"Gong ",
"Pan ",
"Hwang ",
"Dao ",
"Myek ",
"Jia ",
"Teng ",
"Hui ",
"Zhong ",
"Shan ",
"Man ",
"Mak ",
"Biao ",
"Guo ",
"Chayk ",
"Mu ",
"Pang ",
"Zhang ",
"Jiong ",
"Chan ",
"Pok ",
"Chi ",
"Hu ",
"Pen ",
"Tang ",
"Phyey ",
"Hei ",
"",
"Mi ",
"Qiao ",
"Chem ",
"Fen ",
"Mong ",
"Pang ",
"Cwu ",
"Mie ",
"Chu ",
"Jie ",
"Hen ",
"Lan ",
"Kan ",
"Phyeng ",
"Nyen ",
"Qian ",
"Pyeng ",
"Pyeng ",
"Hayng ",
"Kan ",
"Yo ",
"Hwan ",
"Yu ",
"Yu ",
"Ki ",
"Em ",
"Pi ",
"Cheng ",
"Ze ",
"Kwang ",
"Cang ",
"Mo ",
"Qing ",
"Pi ",
"Qin ",
"Dun ",
"Sang ",
"Ki ",
"Ya ",
"Bai ",
"Jie ",
"Se ",
"Lu ",
"Wu ",
"",
"Ku ",
"Ying ",
"Ce ",
"Pho ",
"Cem ",
"Ya ",
"Miao ",
"Kyeng ",
"Ci ",
"Pwu ",
"Tong ",
"Pang ",
"Fei ",
"Sang ",
"Yi ",
"Zhi ",
"Tiao ",
"Zhi ",
"Xiu ",
"To ",
"Cwa ",
"Xiao ",
"Tu ",
"Gui ",
"Ko ",
"Pang ",
"Ceng ",
"You ",
"Bu ",
"Ding ",
"Cheng ",
"Lai ",
"Pi ",
"Ji ",
"Am ",
"Se ",
"Kang ",
"Yong ",
"Tuo ",
"Song ",
"Se ",
"Qing ",
"Yu ",
"Yu ",
"Miao ",
"Sou ",
"Chi ",
"Sang ",
"Phyey ",
"Kwu ",
"He ",
"Hui ",
"Liu ",
"Ha ",
"Lyem ",
"Lang ",
"Swu ",
"Zhi ",
"Pou ",
"Qing ",
"Kwu ",
"Kwu ",
"Kun ",
"Ao ",
"Kwak ",
"Lou ",
"Um ",
"Lyo ",
"Dai ",
"Lu ",
"Yi ",
"Cwu ",
"Cen ",
"Tu ",
"Si ",
"Hum ",
"Myo ",
"Chang ",
"Mwu ",
"Phyey ",
"Kwang ",
"Koc ",
"Koy ",
"Bi ",
"Cang ",
"Hay ",
"Lum ",
"Lin ",
"Liao ",
"Lye ",
"",
"Ying ",
"Xian ",
"Ting ",
"Ong ",
"Li ",
"Cheng ",
"Yin ",
"Xun ",
"Yen ",
"Ceng ",
"Di ",
"Po ",
"Ken ",
"Hoy ",
"Nay ",
"Hoy ",
"Gong ",
"Ip ",
],
"x5f": [
"Kai ",
"Pyen ",
"Yi ",
"Ki ",
"Long ",
"Fen ",
"Ju ",
"Kam ",
"Hyek ",
"Zang ",
"Phyey ",
"Ik ",
"Yi ",
"I ",
"San ",
"Sik ",
"Er ",
"Si ",
"Si ",
"Kwung ",
"Co ",
"In ",
"Hu ",
"Pwul ",
"Hong ",
"Wu ",
"Tui ",
"I ",
"Jiang ",
"Ba ",
"Shen ",
"Cey ",
"Zhang ",
"Jue ",
"To ",
"Fu ",
"Di ",
"Mi ",
"Hyen ",
"Ho ",
"Chao ",
"No ",
"Jing ",
"Zhen ",
"Yi ",
"Mi ",
"Quan ",
"Wan ",
"Shao ",
"Yak ",
"Xuan ",
"Jing ",
"Ton ",
"Cang ",
"Jiang ",
"Kang ",
"Peng ",
"Than ",
"Kang ",
"Bi ",
"Phil ",
"She ",
"Than ",
"Jian ",
"Kwu ",
"Sei ",
"Pal ",
"Bi ",
"Kou ",
"Nagi ",
"Pyel ",
"Xiao ",
"Than ",
"Kuo ",
"Kang ",
"Hong ",
"Mi ",
"Kuo ",
"Man ",
"Jue ",
"Ji ",
"Ji ",
"Gui ",
"Tang ",
"Lok ",
"Lu ",
"Tan ",
"Hyey ",
"Chey ",
"Hwi ",
"Hwi ",
"I ",
"Yi ",
"I ",
"I ",
"Huo ",
"Huo ",
"Shan ",
"Hyeng ",
"Wen ",
"Tong ",
"En ",
"En ",
"Wuk ",
"Chi ",
"Chay ",
"Phyo ",
"Co ",
"Pin ",
"Phayng ",
"Yong ",
"Piao ",
"Chang ",
"Yeng ",
"Chi ",
"Chi ",
"Zhuo ",
"Tuo ",
"Ji ",
"Pang ",
"Zhong ",
"Yek ",
"Wang ",
"Che ",
"Phi ",
"Chi ",
"Ling ",
"Pwul ",
"Wang ",
"Ceng ",
"Co ",
"Wang ",
"Kyeng ",
"Tay ",
"Xi ",
"Swun ",
"Hun ",
"Yang ",
"Hoy ",
"Lyul ",
"Hwu ",
"Wa ",
"Cheng ",
"Zhi ",
"Se ",
"Kyeng ",
"To ",
"Cong ",
"",
"Lai ",
"Cong ",
"Tuk ",
"Pay ",
"Sa ",
"",
"Qi ",
"Sang ",
"Zhi ",
"Cong ",
"Zhou ",
"Lay ",
"E ",
"Xie ",
"Cha ",
"Jian ",
"Chi ",
"Jia ",
"Phyen ",
"Hwang ",
"Pok ",
"Swun ",
"Wei ",
"Pang ",
"Yo ",
"Mi ",
"Hyey ",
"Zheng ",
"Piao ",
"Chi ",
"Tek ",
"Cing ",
"Cing ",
"Bie ",
"Tek ",
"Chong ",
"Chel ",
"Jiao ",
"Wei ",
"Yo ",
"Hwi ",
"Mei ",
"Long ",
"Xiang ",
"Bao ",
"Qu ",
"Sim ",
"Ritsushinben ",
"Phil ",
"Yi ",
"Le ",
"Ren ",
"To ",
"Ding ",
"Gai ",
"Ki ",
"In ",
"Ren ",
"Chan ",
"Tham ",
"Te ",
"Thuk ",
"Gan ",
"Qi ",
"Shi ",
"Chon ",
"Ci ",
"Mang ",
"Mang ",
"Xi ",
"Fan ",
"Ung ",
"Chem ",
"Min ",
"Min ",
"Chwung ",
"Chwung ",
"O ",
"Ji ",
"O ",
"Xi ",
"Ye ",
"You ",
"Wan ",
"Cong ",
"Zhong ",
"Khway ",
"Yu ",
"Pyen ",
"Ki ",
"Qi ",
"Cui ",
"Chim ",
"Tai ",
"Tun ",
"Qian ",
"Nyem ",
"Hun ",
"Hyung ",
"Nyu ",
"Wang ",
"Xian ",
"Hun ",
"Kang ",
"Hol ",
"Kai ",
"Pwun ",
],
"x60": [
"Huai ",
"Tai ",
"Song ",
"Wu ",
"Ou ",
"Chang ",
"Chuang ",
"Ju ",
"Yi ",
"Bao ",
"Cho ",
"Min ",
"Pei ",
"Cak ",
"Zen ",
"Ang ",
"Kou ",
"Ban ",
"No ",
"Nao ",
"Ceng ",
"Pha ",
"Pho ",
"Chep ",
"Gu ",
"Ho ",
"Ju ",
"Tal ",
"Lyeng ",
"Sa ",
"Chou ",
"Di ",
"Thay ",
"I ",
"Tu ",
"You ",
"Fu ",
"Kup ",
"Phyeng ",
"Seng ",
"Wen ",
"Ni ",
"Koy ",
"Fu ",
"Xi ",
"Bi ",
"You ",
"Kep ",
"Xuan ",
"Chong ",
"Bing ",
"Hwang ",
"Xu ",
"Chu ",
"Pi ",
"Xi ",
"Xi ",
"Tan ",
"Koraeru ",
"Zong ",
"Dui ",
"",
"Ki ",
"Yi ",
"Chi ",
"Im ",
"Swun ",
"Si ",
"Xi ",
"Lao ",
"Hang ",
"Kwang ",
"Mo ",
"Zhi ",
"Hyep ",
"Lyen ",
"Tiao ",
"Hwang ",
"Die ",
"Hao ",
"Kong ",
"Gui ",
"Hang ",
"Xi ",
"Xiao ",
"Se ",
"Shi ",
"Kua ",
"Qiu ",
"Yang ",
"Ey ",
"Hui ",
"Chi ",
"Kwal ",
"Yi ",
"Hyung ",
"Koy ",
"Lin ",
"Hoy ",
"Ca ",
"Hyul ",
"Chi ",
"Xiang ",
"Nu ",
"Han ",
"Un ",
"Kak ",
"Thong ",
"Nyem ",
"Kong ",
"Quan ",
"Sik ",
"Hup ",
"Yue ",
"Peng ",
"Ken ",
"De ",
"Hyey ",
"E ",
"Kyuu ",
"Tong ",
"Yan ",
"Kai ",
"Ce ",
"Nao ",
"Yun ",
"Mang ",
"Yong ",
"Yong ",
"Yen ",
"Pi ",
"Kon ",
"Cho ",
"Yel ",
"Yu ",
"Yu ",
"Jie ",
"Sil ",
"Zhe ",
"Lin ",
"Cey ",
"Han ",
"Hao ",
"Hyep ",
"Ti ",
"Bu ",
"Up ",
"Qian ",
"Hoy ",
"Huy ",
"Phay ",
"Mwun ",
"Yi ",
"Heng ",
"Song ",
"Cen ",
"Cheng ",
"Kui ",
"Wu ",
"O ",
"Yu ",
"Li ",
"Lyang ",
"Hwan ",
"Chong ",
"Yi ",
"Yel ",
"Li ",
"Nin ",
"Noy ",
"Ak ",
"Que ",
"Xuan ",
"Qian ",
"Wu ",
"Min ",
"Cong ",
"Pi ",
"Pi ",
"Tek ",
"Chwey ",
"Chang ",
"Min ",
"Li ",
"Kyey ",
"Guan ",
"Guan ",
"Hayng ",
"To ",
"Che ",
"Kong ",
"Tian ",
"Lun ",
"Xi ",
"Kan ",
"Kun ",
"Nyek ",
"Ceng ",
"Chwu ",
"Ton ",
"Guo ",
"Chem ",
"Liang ",
"Wan ",
"Yuan ",
"Jin ",
"Ji ",
"Lam ",
"Yu ",
"Hok ",
"He ",
"Kwen ",
"Tan ",
"Chek ",
"Chek ",
"Nie ",
"Mang ",
"Chel ",
"Hol ",
"Hon ",
"Sek ",
"Chang ",
"Xin ",
"Yu ",
"Hyey ",
"Ak ",
"Swa ",
"Chong ",
"Jian ",
"Yong ",
"Dian ",
"Ju ",
"Cham ",
"Cheng ",
"Tek ",
"Bei ",
"Qie ",
"Can ",
"Dan ",
"Guan ",
"Tha ",
"Noy ",
"Wun ",
"Sang ",
"Chwey ",
"Die ",
"Hwang ",
"Cwun ",
"Kyeng ",
"Ya ",
"Seng ",
"Chuk ",
"Phyen ",
"Hun ",
"Zong ",
"Ti ",
],
"x61": [
"Cho ",
"Swu ",
"Bei ",
"Sen ",
"Wei ",
"Ge ",
"Ken ",
"Wei ",
"Yu ",
"Yu ",
"Bi ",
"Xuan ",
"Huan ",
"Min ",
"Phyak ",
"Uy ",
"Mian ",
"Yong ",
"Kai ",
"Dang ",
"Um ",
"Ak ",
"Chen ",
"Mou ",
"Ke ",
"Ke ",
"Wu ",
"Ay ",
"Hyep ",
"Yan ",
"Nuo ",
"Kam ",
"On ",
"Zong ",
"Sai ",
"Leng ",
"Fen ",
"",
"Kui ",
"Koy ",
"Que ",
"Gong ",
"Yun ",
"Su ",
"So ",
"Ki ",
"Yao ",
"Song ",
"Hwang ",
"Ji ",
"Gu ",
"Ju ",
"Chang ",
"Ni ",
"Xie ",
"Kay ",
"Zheng ",
"Yong ",
"Cao ",
"Sun ",
"Sin ",
"Bo ",
"Kay ",
"Wen ",
"Xie ",
"Hun ",
"Yong ",
"Yang ",
"Lyul ",
"Sao ",
"To ",
"Un ",
"Ca ",
"Xu ",
"Kyem ",
"Thay ",
"Hwang ",
"On ",
"Shen ",
"Ming ",
"",
"She ",
"Cong ",
"Phyo ",
"Mo ",
"Mo ",
"Guo ",
"Chi ",
"Cham ",
"Cham ",
"Cham ",
"Cui ",
"Min ",
"Thuk ",
"Zhang ",
"Thong ",
"O ",
"Shuang ",
"Man ",
"Kwan ",
"Kak ",
"Zao ",
"Jiu ",
"Hyey ",
"Kay ",
"Lian ",
"Ou ",
"Cong ",
"Jin ",
"Yin ",
"Lye ",
"Shang ",
"Wi ",
"Tan ",
"Man ",
"Kan ",
"Sup ",
"Yong ",
"Kyeng ",
"Kang ",
"Di ",
"Zhi ",
"Lwu ",
"Juan ",
"Chek ",
"Chek ",
"Yok ",
"Ping ",
"Liao ",
"Chong ",
"Wu ",
"Yong ",
"Zhi ",
"Tong ",
"Cheng ",
"Qi ",
"Qu ",
"Peng ",
"Pi ",
"Pyel ",
"Chun ",
"Kyo ",
"Cung ",
"Chi ",
"Lyen ",
"Ping ",
"Kwey ",
"Hui ",
"Cho ",
"Cheng ",
"Un ",
"Yin ",
"Huy ",
"Huy ",
"Than ",
"Tan ",
"Duo ",
"Dui ",
"Dui ",
"Su ",
"Jue ",
"Ce ",
"Xiao ",
"Fan ",
"Pwun ",
"Lao ",
"Lao ",
"Tong ",
"Kam ",
"Key ",
"Xian ",
"Min ",
"Kyeng ",
"Liao ",
"Mwu ",
"Cham ",
"Jue ",
"Cu ",
"Hen ",
"Tan ",
"Sheng ",
"Pi ",
"Ek ",
"Chu ",
"Sem ",
"Nao ",
"Tam ",
"Than ",
"Kyeng ",
"Song ",
"Kam ",
"Jiao ",
"Wai ",
"Huan ",
"Dong ",
"Kun ",
"Qin ",
"Qu ",
"Co ",
"Kan ",
"Hay ",
"Ung ",
"O ",
"Mwu ",
"Yek ",
"Lin ",
"Se ",
"Jun ",
"Huai ",
"Men ",
"Lan ",
"Ai ",
"Lin ",
"Yem ",
"Gua ",
"Ha ",
"Chi ",
"Yu ",
"Yin ",
"Dai ",
"Meng ",
"Ai ",
"Mong ",
"Tay ",
"Qi ",
"Mo ",
"Lan ",
"Mwun ",
"Chou ",
"Zhi ",
"Na ",
"Nuo ",
"Yan ",
"Yang ",
"Bo ",
"Zhi ",
"Kuang ",
"Kuang ",
"You ",
"Fu ",
"Liu ",
"Mie ",
"Cing ",
"",
"Chan ",
"Meng ",
"La ",
"Hoy ",
"Hyen ",
"Rang ",
"Cham ",
"Ji ",
"Kwu ",
"Hwan ",
"Sep ",
"Uy ",
],
"x62": [
"Lyen ",
"Nan ",
"Mi ",
"Tang ",
"Jue ",
"Gang ",
"Gang ",
"Tang ",
"Kwa ",
"Yue ",
"Mwu ",
"Jian ",
"Swul ",
"Swu ",
"Yung ",
"Xi ",
"Seng ",
"A ",
"Kyey ",
"Ge ",
"Can ",
"Cang ",
"Hok ",
"Qiang ",
"Zhan ",
"Dong ",
"Chek ",
"Jia ",
"Die ",
"Cek ",
"Al ",
"Kuk ",
"Shi ",
"Kam ",
"Cip ",
"Kui ",
"Gai ",
"Deng ",
"Cen ",
"Chang ",
"Ge ",
"Cen ",
"Cel ",
"Yu ",
"Jian ",
"Yan ",
"Lyuk ",
"Huy ",
"Cen ",
"Huy ",
"Huy ",
"Chak ",
"Tay ",
"Kwu ",
"Ho ",
"Ho ",
"Ho ",
"Ayk ",
"Shi ",
"Lye ",
"Mao ",
"Hu ",
"Lye ",
"Pang ",
"So ",
"Phyen ",
"Dian ",
"Kyeng ",
"Shang ",
"Yi ",
"Yi ",
"Sen ",
"Ho ",
"Pi ",
"Yem ",
"Swu ",
"Shu ",
"Cay ",
"Chal ",
"Qiu ",
"Le ",
"Pok ",
"Pay ",
"Tha ",
"Reng ",
"Pwul ",
"Hameru ",
"Cay ",
"Thak ",
"Zhang ",
"Diao ",
"Kang ",
"Yu ",
"Ku ",
"Han ",
"Shen ",
"Cha ",
"Yi ",
"Gu ",
"Kwu ",
"Wu ",
"Tuo ",
"Qian ",
"Zhi ",
"In ",
"Kuo ",
"Men ",
"Sao ",
"Yang ",
"Niu ",
"Pwun ",
"Cha ",
"Rao ",
"Kup ",
"Qian ",
"Pan ",
"Jia ",
"Yu ",
"Pwu ",
"Ao ",
"Xi ",
"Pi ",
"Ci ",
"Zi ",
"Ayk ",
"Dun ",
"Co ",
"Sung ",
"Ki ",
"Yan ",
"Kuang ",
"Pyen ",
"Cho ",
"Ju ",
"Mwun ",
"Hu ",
"Yue ",
"Kyel ",
"Pha ",
"Qin ",
"Zhen ",
"Zheng ",
"Yun ",
"Wan ",
"Nu ",
"Ek ",
"Se ",
"Co ",
"Pwu ",
"Thwu ",
"Twu ",
"Hang ",
"Cel ",
"Pou ",
"Fu ",
"Pho ",
"Pal ",
"Ao ",
"Thayk ",
"Tuan ",
"Kou ",
"Lun ",
"Qiang ",
"",
"Hu ",
"Bao ",
"Bing ",
"Zhi ",
"Phyeng ",
"Tan ",
"Pu ",
"Phi ",
"Thay ",
"Yao ",
"Zhen ",
"Zha ",
"Yang ",
"Pho ",
"He ",
"Ni ",
"Yi ",
"Ce ",
"Chi ",
"Pi ",
"Za ",
"Mal ",
"Mei ",
"Shen ",
"Ap ",
"Chwu ",
"Qu ",
"Min ",
"Chu ",
"Jia ",
"Pwul ",
"Zhan ",
"Cwu ",
"Tan ",
"Thak ",
"Mwu ",
"Nyem ",
"Lap ",
"Pwu ",
"Pho ",
"Pan ",
"Pak ",
"Ling ",
"Na ",
"Koy ",
"Kyem ",
"Ke ",
"Chek ",
"Pal ",
"Tha ",
"Tha ",
"Yo ",
"Kwu ",
"Col ",
"Pan ",
"Cho ",
"Pay ",
"Bai ",
"Di ",
"Ni ",
"Ke ",
"Hwak ",
"Long ",
"Jian ",
"",
"Yong ",
"Lan ",
"Ning ",
"Pal ",
"Ze ",
"Qian ",
"Hen ",
"Kwal ",
"Sik ",
"Kil ",
"Cung ",
"Nin ",
"Kong ",
"Gong ",
"Kwen ",
"Shuan ",
"Con ",
"Chal ",
"Ko ",
"Chi ",
"Xie ",
"Ce ",
"Hui ",
"Phyeng ",
"Yey ",
"Sup ",
"Na ",
],
"x63": [
"Bo ",
"Ci ",
"Kway ",
"Zhi ",
"Kuo ",
"Duo ",
"Duo ",
"Ci ",
"Sel ",
"An ",
"Nong ",
"Zhen ",
"Kyek ",
"Jiao ",
"Ku ",
"Dong ",
"Na ",
"To ",
"Lyen ",
"Zha ",
"Lu ",
"Die ",
"Wa ",
"Jue ",
"Lyut ",
"Ju ",
"Zhi ",
"Luan ",
"Ya ",
"Zhua ",
"Ta ",
"Xie ",
"Nao ",
"Dang ",
"Jiao ",
"Zheng ",
"Ji ",
"Hui ",
"Xun ",
"Ku ",
"Ay ",
"Tuo ",
"Na ",
"Cwa ",
"Bo ",
"Geng ",
"Ti ",
"Cin ",
"Cheng ",
"Suo ",
"Suo ",
"Keng ",
"Mei ",
"Long ",
"Ju ",
"Peng ",
"Jian ",
"Up ",
"Ceng ",
"Yen ",
"Nuo ",
"Man ",
"Hyep ",
"Cha ",
"Feng ",
"Jiao ",
"O ",
"Kwun ",
"Kwu ",
"Tong ",
"Kon ",
"Huo ",
"Tu ",
"Chak ",
"Pou ",
"Le ",
"Phal ",
"Han ",
"So ",
"Nal ",
"Yen ",
"Ze ",
"Song ",
"Ye ",
"Jue ",
"Pho ",
"Huan ",
"Chek ",
"Zun ",
"Yi ",
"Zhai ",
"Lu ",
"Sou ",
"Tuo ",
"Lao ",
"Sun ",
"Bang ",
"Jian ",
"Hwan ",
"Dao ",
"",
"Wan ",
"Qin ",
"Pong ",
"Sa ",
"Lyel ",
"Min ",
"Mwun ",
"Fu ",
"Bai ",
"Ke ",
"Dao ",
"Wo ",
"Ay ",
"Kwen ",
"Yue ",
"Zong ",
"Chen ",
"Chwu ",
"Chep ",
"Tu ",
"Ben ",
"Nal ",
"Nyem ",
"Nuo ",
"Zu ",
"Wo ",
"Se ",
"Hun ",
"Cheng ",
"Dian ",
"So ",
"Lyun ",
"Qing ",
"Gang ",
"Chel ",
"Swu ",
"To ",
"Pwu ",
"Di ",
"Cang ",
"Gun ",
"Ki ",
"To ",
"Qia ",
"Qi ",
"Pay ",
"Shu ",
"Qian ",
"Ling ",
"Ayk ",
"Ya ",
"Kwul ",
"Zheng ",
"Liang ",
"Kway ",
"Yey ",
"Huo ",
"Shan ",
"Ceng ",
"Lyak ",
"Chay ",
"Tham ",
"Chey ",
"Bing ",
"Cep ",
"Ti ",
"Kong ",
"Chwu ",
"Em ",
"Co ",
"Zou ",
"Kwuk ",
"Tian ",
"Qian ",
"Ken ",
"Bai ",
"Shou ",
"Key ",
"Lu ",
"Guo ",
"Haba ",
"",
"Zhi ",
"Dan ",
"Maang ",
"Xian ",
"So ",
"Guan ",
"Peng ",
"Yen ",
"Nuo ",
"Kan ",
"Zhen ",
"Jiu ",
"Cen ",
"Yu ",
"Yan ",
"Kyu ",
"Nan ",
"Hong ",
"Yu ",
"Pi ",
"Wei ",
"Sai ",
"Zou ",
"Xuan ",
"Myo ",
"Cey ",
"Nie ",
"Sap ",
"Shi ",
"Zong ",
"Zhen ",
"Up ",
"Shun ",
"Heng ",
"Bian ",
"Yang ",
"Hwan ",
"Am ",
"Zuan ",
"An ",
"Se ",
"Al ",
"Ak ",
"Ke ",
"Chwi ",
"Ji ",
"Ti ",
"La ",
"La ",
"Cheng ",
"Kay ",
"Jiu ",
"Chwu ",
"Tu ",
"Key ",
"Hwi ",
"Geng ",
"Chong ",
"Shuo ",
"Sel ",
"Xie ",
"Wen ",
"Ken ",
"Ya ",
"Sap ",
"Zha ",
"Bei ",
"Yao ",
"",
"Dam ",
"Lan ",
"Wen ",
"Qin ",
],
"x64": [
"Chan ",
"Ge ",
"Lou ",
"Zong ",
"Geng ",
"Jiao ",
"Kwu ",
"Qin ",
"Yong ",
"Kak ",
"Chou ",
"Chi ",
"Zhan ",
"Son ",
"Sun ",
"Pak ",
"Hyuk ",
"Rong ",
"Pang ",
"Cha ",
"So ",
"Ke ",
"Yo ",
"To ",
"Zhi ",
"Nu ",
"Xie ",
"Jian ",
"Swu ",
"Qiu ",
"Gao ",
"Xian ",
"Shuo ",
"Sang ",
"Cin ",
"Mie ",
"Ayk ",
"Chwu ",
"Nuo ",
"Shan ",
"Thap ",
"Jie ",
"Tang ",
"Pan ",
"Pan ",
"Thap ",
"Li ",
"To ",
"Kol ",
"Zhi ",
"Wa ",
"Xia ",
"Qian ",
"Wen ",
"Chang ",
"Tian ",
"Zhen ",
"E ",
"Hyu ",
"Nuo ",
"Quan ",
"Cha ",
"Chak ",
"Ge ",
"Wu ",
"En ",
"Sep ",
"Kang ",
"She ",
"Shu ",
"Bai ",
"Yao ",
"Bin ",
"Sou ",
"Tan ",
"Sa ",
"Chan ",
"Suo ",
"Liao ",
"Chong ",
"Chuang ",
"Guo ",
"Pyeng ",
"Feng ",
"Shuai ",
"Di ",
"Qi ",
"",
"Cek ",
"Lian ",
"Tang ",
"Chi ",
"Guan ",
"Lu ",
"Luo ",
"Lou ",
"Chong ",
"Gai ",
"Hu ",
"Zha ",
"Chuang ",
"Tang ",
"Hua ",
"Choy ",
"Nai ",
"Ma ",
"Jiang ",
"Gui ",
"Ying ",
"Chek ",
"Ao ",
"Ci ",
"Nie ",
"Man ",
"Shan ",
"Kwu ",
"Shu ",
"Suo ",
"Tuan ",
"Jiao ",
"Mo ",
"Mo ",
"Cep ",
"Sem ",
"Keng ",
"Phyo ",
"Jiang ",
"Yin ",
"Gou ",
"Qian ",
"Liao ",
"Kyek ",
"Ying ",
"Kwey ",
"Pie ",
"Pie ",
"Lo ",
"Dun ",
"Xian ",
"Ruan ",
"Kui ",
"Zan ",
"Yi ",
"Xun ",
"Thayng ",
"Thayng ",
"Sal ",
"Yo ",
"Heng ",
"Se ",
"Qian ",
"Huang ",
"Thap ",
"Cwun ",
"Nyen ",
"Lin ",
"Zheng ",
"Hwi ",
"Tang ",
"Kyo ",
"Ji ",
"Cao ",
"Dan ",
"Dan ",
"Chel ",
"Pal ",
"Che ",
"Jue ",
"Xiao ",
"Lyo ",
"Ben ",
"Mwu ",
"Qiao ",
"Pha ",
"Chwal ",
"Zhuo ",
"Chan ",
"Tuo ",
"Pak ",
"Qin ",
"Dun ",
"Nian ",
"",
"Xie ",
"Lu ",
"Kyo ",
"Cuan ",
"Tal ",
"Kam ",
"Qiao ",
"Kwa ",
"Kem ",
"Gan ",
"Ong ",
"Lei ",
"Kuo ",
"Lo ",
"Chen ",
"Zhuo ",
"Thayk ",
"Pu ",
"Chak ",
"Kyek ",
"Dang ",
"Suo ",
"Co ",
"Kyeng ",
"Jing ",
"Hwan ",
"Jie ",
"Kum ",
"Kuai ",
"Tam ",
"Hyu ",
"Ge ",
"Pyek ",
"Pyek ",
"Ao ",
"Ke ",
"Ye ",
"",
"Mang ",
"Sou ",
"Mi ",
"Cey ",
"Tay ",
"Thak ",
"To ",
"Xing ",
"Lam ",
"Chal ",
"Ke ",
"Ye ",
"Ru ",
"Ye ",
"Ye ",
"Uy ",
"Hwa ",
"Ji ",
"Pin ",
"Ning ",
"Kak ",
"Chek ",
"Jie ",
"Hwak ",
"Mo ",
"Jian ",
"Xie ",
"Lie ",
"Tan ",
"Pha ",
"Sou ",
"Lu ",
"Lue ",
"Yo ",
"Zhi ",
],
"x65": [
"Pan ",
"Yang ",
"Long ",
"Sa ",
"The ",
"Zan ",
"Nian ",
"Hen ",
"Jun ",
"Huo ",
"Lye ",
"La ",
"Han ",
"Ying ",
"Lu ",
"Long ",
"Qian ",
"Qian ",
"Zan ",
"Qian ",
"Lan ",
"San ",
"Ying ",
"Mei ",
"Yang ",
"Cham ",
"",
"Cuan ",
"Xie ",
"Sep ",
"Luo ",
"Jun ",
"Mi ",
"Li ",
"Chan ",
"Lyen ",
"Than ",
"Zuan ",
"Li ",
"Dian ",
"Wa ",
"Dang ",
"Kyo ",
"Hwak ",
"Lam ",
"Li ",
"Nang ",
"Ci ",
"Gui ",
"Gui ",
"Ki ",
"Xin ",
"Pu ",
"Sui ",
"Swu ",
"Ko ",
"Yu ",
"Kay ",
"Yi ",
"Kong ",
"Gan ",
"Pan ",
"Pang ",
"Ceng ",
"Bo ",
"Dian ",
"Kou ",
"Min ",
"Wu ",
"Ko ",
"He ",
"Ce ",
"Hyo ",
"Mi ",
"Chu ",
"Ge ",
"Di ",
"Se ",
"Kyo ",
"Min ",
"Chen ",
"Kwu ",
"Zhen ",
"Duo ",
"E ",
"Chik ",
"O ",
"Phay ",
"Se ",
"Kyo ",
"Duo ",
"Lian ",
"Nie ",
"Phyey ",
"Chang ",
"Dian ",
"Duo ",
"I ",
"Kam ",
"San ",
"Ke ",
"Yan ",
"Ton ",
"Qi ",
"Dou ",
"Hyo ",
"Duo ",
"Jiao ",
"Kyeng ",
"Yang ",
"Xia ",
"Min ",
"Swu ",
"Ai ",
"Ko ",
"Ai ",
"Ceng ",
"Cek ",
"Zhen ",
"Pwu ",
"Swu ",
"Liao ",
"Qu ",
"Xiong ",
"Xi ",
"Jiao ",
"Sen ",
"Jiao ",
"Zhuo ",
"Twu ",
"Lyem ",
"Phyey ",
"Li ",
"Hyo ",
"Hyo ",
"Mwun ",
"Hak ",
"Qi ",
"Qi ",
"Zhai ",
"Pin ",
"Jue ",
"Zhai ",
"",
"Pi ",
"Pan ",
"Ban ",
"Lan ",
"Yu ",
"Lan ",
"Wei ",
"Twu ",
"Sheng ",
"Lyo ",
"Jia ",
"Kok ",
"Sa ",
"Ka ",
"Yu ",
"Cim ",
"Jiao ",
"Al ",
"Tou ",
"Chu ",
"Kun ",
"Chek ",
"Yin ",
"Pwu ",
"Qiang ",
"Zhan ",
"Qu ",
"Cak ",
"Cham ",
"Tan ",
"Zhuo ",
"Sa ",
"Sin ",
"Cak ",
"Chak ",
"Qin ",
"Lin ",
"Zhuo ",
"Chu ",
"Tan ",
"Zhu ",
"Pang ",
"Xie ",
"Hang ",
"E ",
"Si ",
"Pei ",
"Yu ",
"Mye ",
"Pang ",
"Ki ",
"Cen ",
"Mo ",
"Lye ",
"Phay ",
"Pi ",
"Liu ",
"Pwu ",
"Pang ",
"Sen ",
"Ceng ",
"Jing ",
"Ni ",
"Cok ",
"Zhao ",
"Yi ",
"Lyu ",
"Shao ",
"Jian ",
"Es ",
"Yi ",
"Ki ",
"Zhi ",
"Fan ",
"Piao ",
"Fan ",
"Zhan ",
"Guai ",
"Sui ",
"Yu ",
"Mwu ",
"Ji ",
"Ki ",
"Ki ",
"Huo ",
"Il ",
"Tan ",
"Kwu ",
"Ci ",
"Co ",
"Xie ",
"Tiao ",
"Swun ",
"Wuk ",
"Xu ",
"Xu ",
"Kan ",
"Han ",
"Tay ",
"Di ",
"Wu ",
"Chan ",
"Shi ",
"Kuang ",
"Yang ",
"Shi ",
"Wang ",
"Min ",
"Min ",
"Ton ",
"Chun ",
"O ",
],
"x66": [
"Yun ",
"Bei ",
"Ang ",
"Chi ",
"Ban ",
"Jie ",
"Kon ",
"Sung ",
"Hu ",
"Pang ",
"Ho ",
"Gui ",
"Chang ",
"Xuan ",
"Myeng ",
"Hon ",
"Pwun ",
"Kum ",
"Hu ",
"Yek ",
"Sek ",
"Hun ",
"Yan ",
"Ze ",
"Fang ",
"Tan ",
"Shen ",
"Ke ",
"Yang ",
"Zan ",
"Pyeng ",
"Seng ",
"Yeng ",
"Hyen ",
"Pei ",
"Zhen ",
"Lyeng ",
"Chwun ",
"Hao ",
"May ",
"Cak ",
"Mo ",
"Bian ",
"Kwu ",
"Hun ",
"So ",
"Zong ",
"Si ",
"Ha ",
"Wuk ",
"Fei ",
"Die ",
"Myo ",
"Nil ",
"Chang ",
"On ",
"Dong ",
"Ai ",
"Pyeng ",
"Ang ",
"Cwu ",
"Long ",
"Xian ",
"Kuang ",
"Tiao ",
"Co ",
"Si ",
"Hwang ",
"Hwang ",
"Xuan ",
"Kui ",
"Xu ",
"Kyo ",
"Cin ",
"Zhi ",
"Cin ",
"Sang ",
"Tong ",
"Hong ",
"An ",
"Gai ",
"Xiang ",
"Shai ",
"Xiao ",
"Ye ",
"Yun ",
"Hui ",
"Han ",
"Han ",
"Cwun ",
"Man ",
"Hyen ",
"Kun ",
"Cwu ",
"Huy ",
"Seng ",
"Sheng ",
"Pho ",
"Cel ",
"Zhe ",
"O ",
"Hwan ",
"Hoy ",
"Ho ",
"Sin ",
"Man ",
"Tian ",
"Thak ",
"Swi ",
"Zhou ",
"Po ",
"Kyeng ",
"Sek ",
"Shan ",
"Yi ",
"Sek ",
"Cheng ",
"Qi ",
"Ceng ",
"Kwi ",
"Ceng ",
"Yi ",
"Ci ",
"Am ",
"Wan ",
"Lin ",
"Liang ",
"Chang ",
"Wang ",
"Hyo ",
"Zan ",
"Hi ",
"Hwen ",
"Xuan ",
"Yi ",
"Ka ",
"Hwun ",
"Hwi ",
"Fu ",
"Min ",
"Kyu ",
"He ",
"Yeng ",
"Du ",
"Wi ",
"Se ",
"Qing ",
"Mao ",
"Nan ",
"Jian ",
"Nan ",
"Am ",
"Yang ",
"Chun ",
"Yao ",
"Suo ",
"Jin ",
"Myeng ",
"Kyo ",
"Kai ",
"Ko ",
"Weng ",
"Chang ",
"Ki ",
"Hao ",
"Yan ",
"Li ",
"Ai ",
"Ki ",
"Gui ",
"Men ",
"Cam ",
"Sel ",
"Hao ",
"Mo ",
"Mo ",
"Cong ",
"Nil ",
"Cang ",
"Hye ",
"Phok ",
"Han ",
"Xuan ",
"Chuan ",
"Lyo ",
"Sem ",
"Dan ",
"Kyeng ",
"Pie ",
"Lin ",
"Ton ",
"Huy ",
"Ey ",
"Ki ",
"Huang ",
"Tai ",
"Yep ",
"Yep ",
"Lyek ",
"Tam ",
"Tong ",
"Hyo ",
"Fei ",
"Qin ",
"Zhao ",
"Hao ",
"Yi ",
"Xiang ",
"Xing ",
"Sen ",
"Jiao ",
"Pho ",
"Jing ",
"Yan ",
"Ay ",
"Ye ",
"Ru ",
"Se ",
"Mong ",
"Hwun ",
"Yo ",
"Pho ",
"Li ",
"Chen ",
"Kwang ",
"Cep ",
"",
"Yen ",
"Huo ",
"Lu ",
"Huy ",
"Rong ",
"Long ",
"Nang ",
"La ",
"Luan ",
"Swa ",
"Tang ",
"Em ",
"Chok ",
"Wal ",
"Yue ",
"Kok ",
"Yey ",
"Kyeng ",
"Ye ",
"Hu ",
"Kal ",
"Se ",
"Co ",
"Co ",
"Shou ",
"Man ",
"Ceng ",
"Cung ",
"Chey ",
],
"x67": [
"Choy ",
"Can ",
"Xu ",
"Hoy ",
"Yin ",
"Kel ",
"Fen ",
"Pi ",
"Wel ",
"Yu ",
"Ruan ",
"Pwung ",
"Pan ",
"Pok ",
"Ling ",
"Fei ",
"Qu ",
"",
"Nu ",
"Tiao ",
"Sak ",
"Cim ",
"Lang ",
"Lang ",
"Juan ",
"Ming ",
"Huang ",
"Mang ",
"Tun ",
"Co ",
"Ki ",
"Ki ",
"Ying ",
"Zong ",
"Wang ",
"Tong ",
"Lang ",
"",
"Mong ",
"Long ",
"Mok ",
"Tung ",
"Mi ",
"Mal ",
"Pon ",
"Chal ",
"Chwul ",
"Chwul ",
"",
"Cwu ",
"Ren ",
"Pha ",
"Pak ",
"Tha ",
"Tha ",
"Dao ",
"Li ",
"Qiu ",
"Kwey ",
"Jiu ",
"Bi ",
"Hwu ",
"Ting ",
"Ci ",
"Sha ",
"Eburi ",
"Za ",
"Quan ",
"Qian ",
"Yu ",
"Kan ",
"Wu ",
"Cha ",
"Sam ",
"Xun ",
"Fan ",
"Wu ",
"Zi ",
"Li ",
"Hayng ",
"Cay ",
"Chon ",
"Ren ",
"Phyo ",
"Tuo ",
"Di ",
"Cang ",
"Mang ",
"Chi ",
"Yi ",
"Gu ",
"Gong ",
"Twu ",
"Yi ",
"Ki ",
"Sok ",
"Kang ",
"Co ",
"Moku ",
"Son ",
"Tochi ",
"Lai ",
"Sugi ",
"Mang ",
"Yang ",
"Ma ",
"Cho ",
"Si ",
"Wen ",
"Hang ",
"Phyey ",
"Pay ",
"Kel ",
"Tong ",
"Ko ",
"Myo ",
"Xian ",
"Ce ",
"Chun ",
"Pha ",
"Shu ",
"Hua ",
"Xin ",
"Nyu ",
"Ce ",
"Chou ",
"Song ",
"Phan ",
"Song ",
"Ji ",
"Yue ",
"Jin ",
"Gou ",
"Ji ",
"Mao ",
"Pi ",
"Bi ",
"Wang ",
"Ang ",
"Pang ",
"Pwun ",
"Yi ",
"Fu ",
"Nam ",
"Sek ",
"Hu ",
"Ya ",
"Twu ",
"Xun ",
"Chim ",
"Yao ",
"Lim ",
"Yey ",
"Wa ",
"May ",
"Zhao ",
"Kwa ",
"Ci ",
"Cong ",
"Yun ",
"Waku ",
"Dou ",
"Shu ",
"Zao ",
"",
"Li ",
"Ro ",
"Jian ",
"Cheng ",
"Shou ",
"Qiang ",
"Feng ",
"Nan ",
"Xiao ",
"Xian ",
"Ko ",
"Phyeng ",
"Tay ",
"Si ",
"Ci ",
"Guai ",
"Xiao ",
"Ka ",
"Ka ",
"Kwu ",
"Fu ",
"Mo ",
"Yi ",
"Ye ",
"Ye ",
"Si ",
"Nie ",
"Bi ",
"Tha ",
"Yi ",
"Ling ",
"Pyeng ",
"Ni ",
"La ",
"He ",
"Pan ",
"Fan ",
"Cong ",
"Dai ",
"Ci ",
"Yang ",
"Fu ",
"Payk ",
"Mo ",
"Kam ",
"Chil ",
"Yem ",
"Yu ",
"Mao ",
"Zhao ",
"Song ",
"Ca ",
"Hap ",
"Yu ",
"Shen ",
"Gui ",
"Thak ",
"Cak ",
"Nam ",
"Ning ",
"Yong ",
"Ce ",
"Zhi ",
"Zha ",
"Sa ",
"Dan ",
"Gu ",
"",
"Kwu ",
"Ao ",
"Fu ",
"Kan ",
"Bo ",
"Duo ",
"Ka ",
"Nay ",
"Cwu ",
"Bi ",
"Lyu ",
"Si ",
"Chayk ",
"Sa ",
"Chwuk ",
"Pei ",
"Si ",
"Guai ",
"Sa ",
"Yao ",
"Jue ",
"Kwu ",
"Si ",
],
"x68": [
"Zhi ",
"Liu ",
"Mei ",
"Hoy ",
"Yeng ",
"Chayk ",
"",
"Biao ",
"Zhan ",
"Jie ",
"Long ",
"Dong ",
"Lu ",
"Sayng ",
"Li ",
"Lan ",
"Yong ",
"Shu ",
"Swun ",
"Cen ",
"Qi ",
"Zhen ",
"Se ",
"Lyul ",
"Yi ",
"Xiang ",
"Zhen ",
"Li ",
"Su ",
"Kwal ",
"Kan ",
"Bing ",
"Ren ",
"Kyo ",
"Payk ",
"Ren ",
"Bing ",
"Zi ",
"Chou ",
"Yi ",
"Jie ",
"Xu ",
"Cwu ",
"Jian ",
"Zui ",
"Er ",
"I ",
"Wuk ",
"Fa ",
"Kong ",
"Ko ",
"Lao ",
"Cen ",
"Li ",
"",
"Yang ",
"Hayk ",
"Kun ",
"Zhi ",
"Sik ",
"Kyek ",
"Cay ",
"Luan ",
"Fu ",
"Kel ",
"Hang ",
"Kyey ",
"To ",
"Kwang ",
"Wei ",
"Kwang ",
"Ru ",
"An ",
"An ",
"Juan ",
"Yi ",
"Zhuo ",
"Ku ",
"Cil ",
"Qiong ",
"Tong ",
"Sang ",
"Sang ",
"Hwan ",
"Kil ",
"Jiu ",
"Xue ",
"Duo ",
"Zhui ",
"Yu ",
"Zan ",
"Kasei ",
"Ayng ",
"Masu ",
"",
"Zhan ",
"Ya ",
"Nao ",
"Zhen ",
"Dang ",
"Qi ",
"Qiao ",
"Hua ",
"Hoy ",
"Jiang ",
"Zhuang ",
"Xun ",
"Suo ",
"Sha ",
"Cin ",
"Bei ",
"Ceng ",
"Gua ",
"Jing ",
"Bo ",
"Ben ",
"Pwu ",
"Rui ",
"Thong ",
"Kak ",
"Xi ",
"Lang ",
"Liu ",
"Feng ",
"Qi ",
"Wen ",
"Kwun ",
"Han ",
"Cu ",
"Lyang ",
"Qiu ",
"Ceng ",
"You ",
"May ",
"Pang ",
"Long ",
"Peng ",
"Zhuang ",
"Di ",
"Xuan ",
"Tu ",
"Zao ",
"Ao ",
"Kok ",
"Bi ",
"Di ",
"Han ",
"Cay ",
"Chi ",
"Ren ",
"Bei ",
"Kyeng ",
"Jian ",
"Huan ",
"Wan ",
"Na ",
"Jia ",
"Co ",
"Ji ",
"Hyo ",
"Lye ",
"Wan ",
"Cho ",
"Sim ",
"Fen ",
"Song ",
"Meng ",
"O ",
"Li ",
"Li ",
"Dou ",
"Cen ",
"Ying ",
"Sa ",
"Ju ",
"Cey ",
"Kyey ",
"Kon ",
"Zhuo ",
"So ",
"Chan ",
"Pem ",
"Mi ",
"Jing ",
"Li ",
"Bing ",
"Fumoto ",
"Shikimi ",
"Tao ",
"Zhi ",
"Lai ",
"Lian ",
"Jian ",
"Zhuo ",
"Ling ",
"Li ",
"Ki ",
"Pyeng ",
"Zhun ",
"Cong ",
"Qian ",
"Myen ",
"Ki ",
"Ki ",
"Chay ",
"Kon ",
"Chan ",
"Te ",
"Pi ",
"Pai ",
"Pong ",
"Pou ",
"Hun ",
"Cong ",
"Ceng ",
"Co ",
"Kuk ",
"Li ",
"Pwung ",
"Yu ",
"Yu ",
"Gu ",
"Hun ",
"Tong ",
"Tang ",
"Gang ",
"Wang ",
"Chey ",
"Xi ",
"Fan ",
"Cheng ",
"Can ",
"Kyey ",
"Yuan ",
"Yan ",
"Yu ",
"Kwen ",
"Yi ",
"Sam ",
"Ren ",
"Chui ",
"Leng ",
"Se ",
"Zhuo ",
"Fu ",
"Ke ",
"Lai ",
"Zou ",
"Zou ",
"To ",
"Kwan ",
"Pwun ",
"Pwun ",
"Chen ",
"Qiong ",
"Nie ",
],
"x69": [
"Wan ",
"Kwak ",
"Lu ",
"Hao ",
"Cep ",
"Uy ",
"Chou ",
"Ju ",
"Ju ",
"Cheng ",
"Zuo ",
"Lyang ",
"Qiang ",
"Sik ",
"Chwu ",
"Ya ",
"Ju ",
"Pi ",
"Cho ",
"Zhuo ",
"Zi ",
"Bin ",
"Peng ",
"Ding ",
"Chu ",
"Chang ",
"Mun ",
"Momiji ",
"Kem ",
"Gui ",
"Xi ",
"Du ",
"Qian ",
"Kunugi ",
"Kai ",
"Shide ",
"Luo ",
"Zhi ",
"Ken ",
"Myeng ",
"Tafu ",
"",
"Peng ",
"Zhan ",
"",
"Tuo ",
"Sen ",
"Duo ",
"Ya ",
"Fou ",
"Wei ",
"Wei ",
"Tan ",
"Jia ",
"Cong ",
"Jian ",
"Yi ",
"Chim ",
"Po ",
"Yan ",
"Yan ",
"Yen ",
"Zhan ",
"Chwun ",
"Yu ",
"He ",
"Sa ",
"Wo ",
"Pian ",
"Bi ",
"Yao ",
"Huo ",
"Xu ",
"Ruo ",
"Yang ",
"La ",
"Yan ",
"Ben ",
"Hun ",
"Kyu ",
"Jie ",
"Kui ",
"Si ",
"Phwung ",
"Sel ",
"Tha ",
"Zhi ",
"Ken ",
"Mu ",
"Mwu ",
"Cho ",
"Hu ",
"Hu ",
"Lyen ",
"Lung ",
"Ting ",
"Nam ",
"Yu ",
"Yu ",
"Mi ",
"Song ",
"Wen ",
"Wen ",
"Ying ",
"Ceng ",
"Pian ",
"Cep ",
"Cup ",
"Kal ",
"Ep ",
"Ce ",
"Swun ",
"Yu ",
"Cou ",
"Wei ",
"May ",
"Di ",
"Kuk ",
"Jie ",
"Hay ",
"Chwu ",
"Yeng ",
"Rou ",
"Heng ",
"Lwu ",
"Le ",
"Hazou ",
"Katsura ",
"Pin ",
"Muro ",
"Kay ",
"Tan ",
"Lan ",
"Yun ",
"Yu ",
"Chen ",
"Lu ",
"Ju ",
"Sakaki ",
"",
"Pi ",
"Xie ",
"Ka ",
"Yi ",
"Zhan ",
"Pwu ",
"Nai ",
"Mi ",
"Lang ",
"Yong ",
"Kok ",
"Jian ",
"Kwu ",
"Ta ",
"Yao ",
"Cin ",
"Pang ",
"Sha ",
"Yuan ",
"Cay ",
"Ming ",
"Su ",
"Jia ",
"Yao ",
"Kel ",
"Hwang ",
"Kan ",
"Pi ",
"Ca ",
"Qian ",
"Ma ",
"Sun ",
"Yuan ",
"Sa ",
"Yeng ",
"Shi ",
"Ci ",
"Choy ",
"Yun ",
"Ting ",
"Lyu ",
"Rong ",
"Tang ",
"Kyo ",
"Zhai ",
"Si ",
"Sheng ",
"Thap ",
"Hap ",
"Xi ",
"Kol ",
"Qi ",
"Ko ",
"Ko ",
"Sun ",
"Pan ",
"Tao ",
"Ge ",
"Xun ",
"Cen ",
"Nou ",
"Ji ",
"Sak ",
"Kwu ",
"Thoy ",
"Chang ",
"Cha ",
"Qian ",
"Koy ",
"Mei ",
"Xu ",
"Kong ",
"Gao ",
"Zhuo ",
"Tuo ",
"Kyou ",
"Yang ",
"Cen ",
"Jia ",
"Jian ",
"Zui ",
"Tou ",
"Rou ",
"Bin ",
"Zhu ",
"",
"Xi ",
"Qi ",
"Lian ",
"Hyey ",
"Yong ",
"Cham ",
"Kwak ",
"Kay ",
"Kay ",
"Tuan ",
"Hua ",
"Sayk ",
"Sen ",
"Cui ",
"Beng ",
"You ",
"Kok ",
"Jiang ",
"Hu ",
"Huan ",
"Kui ",
"Yi ",
"Nie ",
"Ko ",
"Kang ",
"Kyu ",
"Gui ",
"Co ",
"Man ",
"Kun ",
],
"x6a": [
"Cek ",
"Zhuang ",
"Ak ",
"Lang ",
"Chen ",
"Cong ",
"Li ",
"Xiu ",
"Qing ",
"Shuang ",
"Pen ",
"Thong ",
"Guan ",
"Ji ",
"Suo ",
"Lei ",
"Lu ",
"Lyang ",
"Mil ",
"Lwu ",
"So ",
"Su ",
"Ke ",
"Ce ",
"Tang ",
"Phyo ",
"Lu ",
"Kyu ",
"Shu ",
"Zha ",
"Chwu ",
"Cang ",
"Men ",
"Mo ",
"Niao ",
"Yang ",
"Tiao ",
"Peng ",
"Zhu ",
"Sha ",
"Chi ",
"Kwen ",
"Hoyng ",
"Kyen ",
"Cong ",
"",
"Hokuso ",
"Qiang ",
"Tara ",
"Ying ",
"Er ",
"Xin ",
"Zhi ",
"Cho ",
"Zui ",
"Cong ",
"Pok ",
"Swu ",
"Hwa ",
"Kwey ",
"Zhen ",
"Cwun ",
"Yue ",
"Zhan ",
"Xi ",
"Xun ",
"Dian ",
"Pel ",
"Kam ",
"Mo ",
"Wu ",
"Chwi ",
"Yo ",
"Lin ",
"Lyu ",
"Kyo ",
"Xian ",
"Run ",
"Fan ",
"Zhan ",
"Thak ",
"Lao ",
"Wun ",
"Swun ",
"Tui ",
"Cheng ",
"Tang ",
"Meng ",
"Kyul ",
"Tung ",
"Swuk ",
"Jue ",
"Jue ",
"Tan ",
"Hui ",
"Ki ",
"Nuo ",
"Sang ",
"Tha ",
"Ning ",
"Rui ",
"Zhu ",
"Tong ",
"Zeng ",
"Fen ",
"Qiong ",
"Ran ",
"Hoyng ",
"Cen ",
"Gu ",
"Liu ",
"Lao ",
"Gao ",
"Chu ",
"Zusa ",
"Nude ",
"Ca ",
"San ",
"Ji ",
"Dou ",
"Shou ",
"Lu ",
"Gian ",
"",
"Yuan ",
"Ta ",
"Shu ",
"Kang ",
"Tan ",
"Lin ",
"Nong ",
"Yin ",
"Kyek ",
"Sui ",
"Shan ",
"Zui ",
"Xuan ",
"Ceng ",
"Gan ",
"Ju ",
"Zui ",
"Ek ",
"Kum ",
"Pu ",
"Chem ",
"Lei ",
"Feng ",
"Hui ",
"Tang ",
"Ji ",
"Sui ",
"Pyek ",
"Bi ",
"Ding ",
"Chu ",
"Zhua ",
"Hoy ",
"Cup ",
"Jie ",
"Ka ",
"Kyeng ",
"Zhe ",
"Kem ",
"Cang ",
"Dao ",
"Yi ",
"Biao ",
"Song ",
"She ",
"Lin ",
"Reki ",
"Cha ",
"Meng ",
"Yin ",
"To ",
"Tay ",
"Mian ",
"Qi ",
"",
"Pin ",
"Huo ",
"Ji ",
"Yem ",
"Mi ",
"Ning ",
"Yi ",
"Gao ",
"Ham ",
"Un ",
"Er ",
"Qing ",
"Yem ",
"Qi ",
"Mi ",
"To ",
"Kwey ",
"Chun ",
"Ji ",
"Kui ",
"Po ",
"Deng ",
"Chu ",
"Han ",
"Mian ",
"You ",
"Zhi ",
"Guang ",
"Qian ",
"Lei ",
"Lei ",
"Sa ",
"Lo ",
"Li ",
"Cuan ",
"Lu ",
"Mie ",
"Hui ",
"Ou ",
"Lye ",
"Cul ",
"Ko ",
"Tok ",
"Yen ",
"Lyek ",
"Fei ",
"Zhuo ",
"Sou ",
"Lian ",
"Sen ",
"Chu ",
"",
"Zhu ",
"Lu ",
"Yan ",
"Lyek ",
"Zhu ",
"Chin ",
"Jie ",
"E ",
"Su ",
"Huai ",
"Nie ",
"Yu ",
"Long ",
"Lai ",
"",
"Hen ",
"Kwi ",
"Ju ",
"Xiao ",
"Ling ",
"Ayng ",
"Chem ",
"Yin ",
"You ",
"Ying ",
],
"x6b": [
"Xiang ",
"Nong ",
"Pak ",
"Cham ",
"Lan ",
"Ju ",
"Shuang ",
"She ",
"Wei ",
"Cong ",
"Kwen ",
"Qu ",
"Cang ",
"",
"Yu ",
"Luo ",
"Li ",
"Chan ",
"Lan ",
"Dang ",
"Jue ",
"Em ",
"Lam ",
"Lan ",
"Zhu ",
"Lei ",
"Li ",
"Ba ",
"Nang ",
"Yu ",
"Lyeng ",
"Tsuki ",
"Hum ",
"Cha ",
"Huan ",
"Hun ",
"Yu ",
"Yu ",
"Qian ",
"Ou ",
"Xu ",
"Chao ",
"Chu ",
"Chi ",
"Hay ",
"Yi ",
"Jue ",
"Xi ",
"Xu ",
"Xia ",
"Yok ",
"Kuai ",
"Lang ",
"Kuan ",
"Shuo ",
"Huy ",
"Ay ",
"Uy ",
"Ki ",
"Hwul ",
"Chi ",
"Hum ",
"Kwan ",
"Kam ",
"Kuan ",
"Kan ",
"Chuan ",
"Sap ",
"",
"Yin ",
"Hum ",
"Hel ",
"Yu ",
"Kyem ",
"Xiao ",
"Yi ",
"Ka ",
"Wu ",
"Than ",
"Jin ",
"Kwu ",
"Hu ",
"Ti ",
"Hwan ",
"He ",
"Pen ",
"Xi ",
"Xiao ",
"Xu ",
"Hup ",
"Sen ",
"Kem ",
"Chu ",
"Yi ",
"Kan ",
"Ye ",
"Chel ",
"Hwan ",
"Ci ",
"Ceng ",
"Cha ",
"Po ",
"Mwu ",
"Ki ",
"Bu ",
"Po ",
"Way ",
"Ju ",
"Qian ",
"Chi ",
"Se ",
"Chi ",
"Se ",
"Zhong ",
"Sey ",
"Sey ",
"Li ",
"Cuo ",
"Yu ",
"Lyek ",
"Kwi ",
"Al ",
"Dai ",
"Sa ",
"Jian ",
"Zhe ",
"Mo ",
"Mol ",
"Yo ",
"Mol ",
"Co ",
"Ang ",
"Cin ",
"Sheng ",
"Thay ",
"Shang ",
"Xu ",
"Swun ",
"Swu ",
"Can ",
"Jue ",
"Phyo ",
"Qia ",
"Qiu ",
"Su ",
"Kung ",
"Yun ",
"Lian ",
"Yi ",
"Fou ",
"Sik ",
"Ye ",
"Can ",
"Hun ",
"Dan ",
"Ji ",
"Ye ",
"",
"Wun ",
"Wen ",
"Chou ",
"Bin ",
"Chey ",
"Jin ",
"Sang ",
"Yin ",
"Diao ",
"Cu ",
"Hui ",
"Cuan ",
"Yi ",
"Than ",
"Du ",
"Kang ",
"Lyem ",
"Pin ",
"Du ",
"Sen ",
"Sem ",
"Swi ",
"Ou ",
"Tan ",
"Zhu ",
"Un ",
"Qing ",
"Yi ",
"Sal ",
"Kak ",
"Kak ",
"Hyo ",
"Jun ",
"Cen ",
"Hwey ",
"Hwey ",
"Gu ",
"Que ",
"Kyek ",
"Uy ",
"Kwu ",
"Hui ",
"Duan ",
"Yi ",
"Xiao ",
"Mwu ",
"Guan ",
"Mo ",
"May ",
"May ",
"Ai ",
"Zuo ",
"Tok ",
"Yuk ",
"Pi ",
"Bi ",
"Pi ",
"Pi ",
"Pi ",
"Bi ",
"Cham ",
"Mo ",
"",
"",
"Pu ",
"Mushiru ",
"Jia ",
"Zhan ",
"Sai ",
"Mu ",
"Tuo ",
"Xun ",
"Er ",
"Rong ",
"Xian ",
"Ju ",
"Mu ",
"Ho ",
"Kwu ",
"Dou ",
"Mushiru ",
"Tam ",
"Pei ",
"Ju ",
"Duo ",
"Chwi ",
"Bi ",
"San ",
"",
"Mao ",
"Sui ",
"Shu ",
"Yu ",
"Tuo ",
"He ",
"Jian ",
"Ta ",
"Sam ",
],
"x6c": [
"Lu ",
"Mu ",
"Mao ",
"Tong ",
"Rong ",
"Chang ",
"Pu ",
"Luo ",
"Cen ",
"Sao ",
"Zhan ",
"Meng ",
"Luo ",
"Qu ",
"Die ",
"Ssi ",
"Ce ",
"Min ",
"Jue ",
"Mayng ",
"Qi ",
"Pie ",
"Nai ",
"Ki ",
"Dao ",
"Xian ",
"Chuan ",
"Pwun ",
"Ri ",
"Nei ",
"",
"Fu ",
"Shen ",
"Dong ",
"Qing ",
"Ki ",
"In ",
"Xi ",
"Hai ",
"Yang ",
"An ",
"Ya ",
"Ke ",
"Qing ",
"Ya ",
"Dong ",
"Dan ",
"Lu ",
"Qing ",
"Yang ",
"On ",
"On ",
"Swu ",
"Sanzui ",
"Zheng ",
"Ping ",
"Yeng ",
"Dang ",
"Sui ",
"Le ",
"Ni ",
"Than ",
"Pem ",
"Kwey ",
"Ceng ",
"Cip ",
"Kwu ",
"Pha ",
"Ze ",
"Mian ",
"Cuan ",
"Hui ",
"Diao ",
"Yi ",
"Cha ",
"Cak ",
"Chuan ",
"Wan ",
"Pem ",
"Tay ",
"Sek ",
"Tuo ",
"Mang ",
"Qiu ",
"Hul ",
"San ",
"Pai ",
"Han ",
"Qian ",
"O ",
"O ",
"Sin ",
"Sa ",
"Ye ",
"Hong ",
"Kang ",
"Ci ",
"O ",
"Tsuchi ",
"",
"Tang ",
"Zhi ",
"Chi ",
"Qian ",
"Myek ",
"Yul ",
"Wang ",
"Qing ",
"Jing ",
"Yey ",
"Jun ",
"Hong ",
"Thay ",
"Quan ",
"Kup ",
"Bian ",
"Pha ",
"Gan ",
"Mwun ",
"Zhong ",
"Fang ",
"Xiong ",
"Kyel ",
"Hang ",
"",
"Ki ",
"Pwun ",
"Xu ",
"Xu ",
"Sim ",
"Ki ",
"Ok ",
"Wun ",
"Wen ",
"Hang ",
"Yen ",
"Chim ",
"Chim ",
"Dan ",
"You ",
"Ton ",
"Ho ",
"Huo ",
"Qi ",
"Mok ",
"Rou ",
"Mol ",
"Tap ",
"Myen ",
"Mwul ",
"Chwung ",
"Tian ",
"Pi ",
"Sa ",
"Ci ",
"Phay ",
"Pan ",
"Zhui ",
"Za ",
"Gou ",
"Liu ",
"Mei ",
"Thayk ",
"Feng ",
"Ou ",
"Li ",
"Lun ",
"Cang ",
"Feng ",
"Wei ",
"Hu ",
"Mal ",
"May ",
"Shu ",
"Ce ",
"Zan ",
"Thak ",
"Tha ",
"Tuo ",
"Ha ",
"Li ",
"Mi ",
"Yi ",
"Fa ",
"Pi ",
"Yu ",
"Tian ",
"Chi ",
"So ",
"Ko ",
"Chem ",
"Yen ",
"Si ",
"Hwang ",
"Hyeng ",
"Ju ",
"Sel ",
"Swi ",
"Il ",
"Jia ",
"Zhong ",
"Chen ",
"Pak ",
"Hui ",
"Phil ",
"Ben ",
"Zhuo ",
"Chu ",
"Luk ",
"Yu ",
"Gu ",
"Hong ",
"Kem ",
"Pep ",
"Mao ",
"Sa ",
"Hu ",
"Phyeng ",
"Ca ",
"Pem ",
"Ci ",
"So ",
"Ning ",
"Cheng ",
"Lyeng ",
"Pho ",
"Pha ",
"Up ",
"Si ",
"Ni ",
"Ju ",
"Yue ",
"Cwu ",
"Sheng ",
"Lei ",
"Hyen ",
"Xue ",
"Fu ",
"Pan ",
"Min ",
"Thay ",
"Ang ",
"Ji ",
"Yeng ",
"Guan ",
"Beng ",
"Xue ",
"Long ",
"Lu ",
"",
"Bo ",
"Xie ",
"Po ",
"Ze ",
"Jing ",
"Yin ",
],
"x6d": [
"Zhou ",
"Ji ",
"Yi ",
"Hui ",
"Hoy ",
"Zui ",
"Cheng ",
"In ",
"Wei ",
"Hou ",
"Chen ",
"Yang ",
"Lyel ",
"Si ",
"Ji ",
"Er ",
"Xing ",
"Po ",
"Sey ",
"Suo ",
"Zhi ",
"Yin ",
"Wu ",
"Sey ",
"Kao ",
"Swu ",
"Jiang ",
"Lak ",
"",
"An ",
"Tong ",
"I ",
"Mou ",
"Lei ",
"Yi ",
"Mi ",
"Quan ",
"Cin ",
"Mo ",
"Yu ",
"Xiao ",
"Sel ",
"Hong ",
"Hyek ",
"Sok ",
"Kuang ",
"To ",
"Qie ",
"Ju ",
"Er ",
"Cwu ",
"Ru ",
"Ping ",
"Swun ",
"Hyung ",
"Zhi ",
"Kwang ",
"Wen ",
"Myeng ",
"Hwal ",
"Wa ",
"Hup ",
"Pha ",
"Wu ",
"Qu ",
"Lyu ",
"Yi ",
"Jia ",
"Jing ",
"Chen ",
"Jiang ",
"Jiao ",
"Cheng ",
"Shi ",
"Zhuo ",
"Ce ",
"Pal ",
"Kuai ",
"Ji ",
"Liu ",
"Chan ",
"Hun ",
"Hu ",
"Nong ",
"Xun ",
"Jin ",
"Lie ",
"Qiu ",
"Wei ",
"Cel ",
"Cwun ",
"Han ",
"Pin ",
"Mang ",
"Zhuo ",
"You ",
"Xi ",
"Pal ",
"Dou ",
"Wan ",
"Koyng ",
"Up ",
"Pho ",
"Yeng ",
"Lan ",
"Ho ",
"Lang ",
"Han ",
"Li ",
"Geng ",
"Pwu ",
"Wu ",
"Lian ",
"Chun ",
"Feng ",
"Yi ",
"Yok ",
"Tong ",
"Lao ",
"Hay ",
"Chim ",
"Hyep ",
"Chong ",
"Weng ",
"Mei ",
"Sui ",
"Cheng ",
"Phay ",
"Xian ",
"Shen ",
"To ",
"Kun ",
"Pin ",
"Nyel ",
"Han ",
"Kyeng ",
"So ",
"Sep ",
"Nian ",
"Tu ",
"Yong ",
"Hyo ",
"Yen ",
"Ting ",
"E ",
"So ",
"Tun ",
"Yen ",
"Cam ",
"Chey ",
"Li ",
"Shui ",
"Si ",
"Lwu ",
"Shui ",
"Tao ",
"Tok ",
"Lao ",
"Lai ",
"Lian ",
"Wei ",
"Wo ",
"Yun ",
"Hwan ",
"Di ",
"",
"Run ",
"Jian ",
"Zhang ",
"Se ",
"Pwu ",
"Kwan ",
"Xing ",
"Shou ",
"Shuan ",
"Ay ",
"Chuo ",
"Zhang ",
"Ayk ",
"Kong ",
"Wan ",
"Ham ",
"Tuo ",
"Tong ",
"Hwu ",
"Wo ",
"Ju ",
"Gan ",
"Lyang ",
"Hun ",
"Ta ",
"Thak ",
"Ceng ",
"Qie ",
"De ",
"Kwen ",
"Chi ",
"Sek ",
"Hyo ",
"Ki ",
"Kwul ",
"Guo ",
"Han ",
"Lim ",
"Tang ",
"Zhou ",
"Peng ",
"Ho ",
"Chang ",
"Swuk ",
"Che ",
"Fang ",
"Chi ",
"Lu ",
"Nyo ",
"Ju ",
"To ",
"Cong ",
"Lwu ",
"Zhi ",
"Peng ",
"Pi ",
"Song ",
"Tian ",
"Pi ",
"Tam ",
"Yek ",
"Ni ",
"E ",
"Lok ",
"Kem ",
"Mi ",
"Ceng ",
"Ling ",
"Lyun ",
"Um ",
"Swi ",
"Qu ",
"Hoy ",
"Yu ",
"Nian ",
"Sim ",
"Piao ",
"Swun ",
"Wa ",
"Yen ",
"Lay ",
"Hon ",
"Cheng ",
"Em ",
"Chen ",
"Chem ",
"Myo ",
"Zhi ",
"Yin ",
"Mi ",
],
"x6e": [
"Ben ",
"Yuan ",
"Wen ",
"Re ",
"Fei ",
"Cheng ",
"Yuan ",
"Kal ",
"Cey ",
"Sep ",
"Yuan ",
"Juu ",
"Lok ",
"Zi ",
"Du ",
"",
"Jian ",
"Min ",
"Pi ",
"Kei ",
"Yu ",
"Yuan ",
"Shen ",
"Cham ",
"Rou ",
"Hwan ",
"Ce ",
"Kam ",
"Nuan ",
"Thwu ",
"Qiu ",
"Ceng ",
"Ke ",
"To ",
"Feng ",
"Sa ",
"Pal ",
"Ak ",
"Wa ",
"Di ",
"Oy ",
"On ",
"Ru ",
"Sel ",
"Chuk ",
"Wi ",
"Ge ",
"Hang ",
"Yan ",
"Hong ",
"Sen ",
"Mi ",
"Kal ",
"Mao ",
"Yeng ",
"Yen ",
"Yu ",
"Hong ",
"Myo ",
"Xing ",
"Mi ",
"Cay ",
"Hon ",
"Nai ",
"Kui ",
"Shi ",
"E ",
"Pay ",
"Mi ",
"Lyen ",
"Qi ",
"Qi ",
"Mei ",
"Tian ",
"Cwu ",
"Wei ",
"Can ",
"Tan ",
"Myen ",
"Hui ",
"Mo ",
"Se ",
"Ji ",
"Pwun ",
"Cen ",
"Jian ",
"Ho ",
"Feng ",
"Sang ",
"Yi ",
"Yin ",
"Tam ",
"Sik ",
"Jie ",
"Ceng ",
"Hwang ",
"Tan ",
"Yu ",
"Bi ",
"Min ",
"Shi ",
"Tu ",
"Sheng ",
"Yong ",
"Qu ",
"Zhong ",
"",
"Chwu ",
"Jiao ",
"qiu ",
"In ",
"Thang ",
"Long ",
"Huo ",
"Wen ",
"Nam ",
"Ban ",
"You ",
"Quan ",
"Chui ",
"Liang ",
"Chan ",
"Po ",
"Chun ",
"Nie ",
"Zi ",
"Wan ",
"Sup ",
"Man ",
"Yeng ",
"Ratsu ",
"Kui ",
"",
"Jian ",
"Xu ",
"Lu ",
"Gui ",
"Kay ",
"",
"",
"Pal ",
"Jin ",
"Gui ",
"Tang ",
"Wen ",
"Suo ",
"Yuan ",
"Lian ",
"Yao ",
"Mong ",
"Cwun ",
"Sheng ",
"Hap ",
"Tai ",
"Da ",
"Wa ",
"Lyu ",
"Kwu ",
"Sao ",
"Myeng ",
"Zha ",
"Shi ",
"Il ",
"Lun ",
"Ma ",
"Pwu ",
"Wei ",
"Lyul ",
"Cai ",
"Wu ",
"Kyey ",
"On ",
"Qiang ",
"Ze ",
"Shi ",
"So ",
"Yi ",
"Cin ",
"Swu ",
"Yun ",
"Xiu ",
"Un ",
"Yong ",
"Hon ",
"Su ",
"Su ",
"Nik ",
"Ta ",
"Shi ",
"Yok ",
"Wei ",
"Pan ",
"Chu ",
"Ce ",
"Pang ",
"Ong ",
"Chang ",
"Myel ",
"He ",
"Dian ",
"Ho ",
"Hwang ",
"Xi ",
"Ca ",
"Chek ",
"Zhi ",
"Hyeng ",
"Fu ",
"Jie ",
"Hwal ",
"Ge ",
"Cay ",
"To ",
"Tung ",
"Sui ",
"Bi ",
"Jiao ",
"Hoy ",
"Kon ",
"Yin ",
"Gao ",
"Long ",
"Zhi ",
"Yan ",
"She ",
"Man ",
"Ying ",
"Chun ",
"Lu ",
"Lan ",
"Luan ",
"",
"Bin ",
"Tan ",
"Yu ",
"Sou ",
"Ho ",
"Phil ",
"Biao ",
"Chey ",
"Jiang ",
"Kou ",
"Cham ",
"Shang ",
"Cek ",
"Mil ",
"Ao ",
"Lo ",
"Ho ",
"Hu ",
"You ",
"Chan ",
"Fan ",
"Yong ",
"Kon ",
"Man ",
],
"x6f": [
"Qing ",
"E ",
"Phyo ",
"Ji ",
"Ay ",
"Jiao ",
"Chil ",
"Xi ",
"Ji ",
"Lok ",
"Lu ",
"Long ",
"Kun ",
"Guo ",
"Cong ",
"Lwu ",
"Zhi ",
"Kay ",
"Qiang ",
"Li ",
"Yen ",
"Co ",
"Jiao ",
"Cong ",
"Chun ",
"Tan ",
"Kwu ",
"Teng ",
"Ye ",
"Xi ",
"Mi ",
"Tang ",
"Mak ",
"Shang ",
"Han ",
"Lyen ",
"Lan ",
"Wa ",
"Li ",
"Qian ",
"Feng ",
"Xuan ",
"Uy ",
"Man ",
"Ci ",
"Mang ",
"Kang ",
"Ta ",
"Pwul ",
"Swu ",
"Chang ",
"Cang ",
"Chong ",
"Xu ",
"Huan ",
"Kuo ",
"Cem ",
"Yan ",
"Chuang ",
"Liao ",
"Choy ",
"Ti ",
"Yang ",
"Cang ",
"Cong ",
"Yeng ",
"Hong ",
"Xun ",
"Shu ",
"Guan ",
"Ying ",
"Xiao ",
"",
"",
"Xu ",
"Lian ",
"Zhi ",
"Wei ",
"Pi ",
"Yul ",
"Jiao ",
"Pal ",
"Dang ",
"Hui ",
"Kyel ",
"Wu ",
"Pa ",
"Cip ",
"Pan ",
"Kyu ",
"Swu ",
"Cam ",
"Cam ",
"Hup ",
"Lo ",
"Sek ",
"Sun ",
"Ton ",
"Hwang ",
"Min ",
"Lyun ",
"Su ",
"Lyo ",
"Zhen ",
"Zhong ",
"Yi ",
"Di ",
"Wan ",
"Dan ",
"Tam ",
"Co ",
"Sim ",
"Kwey ",
"",
"Shao ",
"Tu ",
"Ce ",
"San ",
"Hei ",
"Bi ",
"San ",
"Chan ",
"Can ",
"Shu ",
"Tong ",
"Po ",
"Lin ",
"Wei ",
"Se ",
"Sap ",
"Cing ",
"Jiong ",
"Cing ",
"Hua ",
"Yo ",
"Lao ",
"Chel ",
"Kem ",
"Cun ",
"Hoyng ",
"Si ",
"Cwu ",
"Phayng ",
"Han ",
"Wun ",
"Liu ",
"Hong ",
"Fu ",
"Ho ",
"He ",
"Xian ",
"Kan ",
"San ",
"Sek ",
"Iku ",
"",
"Lan ",
"",
"Yu ",
"Lum ",
"Mian ",
"Co ",
"Dang ",
"Han ",
"Thayk ",
"Xie ",
"Yu ",
"Lyey ",
"Se ",
"Xue ",
"Ling ",
"Man ",
"Zi ",
"Yong ",
"Hoy ",
"Chan ",
"Lian ",
"Cen ",
"Ye ",
"O ",
"Huan ",
"Zhen ",
"Cen ",
"Man ",
"Dan ",
"Tam ",
"Yi ",
"Sui ",
"Pi ",
"Ju ",
"Tan ",
"Qin ",
"Kyek ",
"Thak ",
"Lyem ",
"Nong ",
"Guo ",
"Jin ",
"Pwun ",
"Sayk ",
"Ji ",
"Swu ",
"Yey ",
"Chu ",
"Ta ",
"Song ",
"Ding ",
"",
"Zhu ",
"Lai ",
"Bin ",
"Lian ",
"Mi ",
"Sup ",
"Shu ",
"Mi ",
"Nyeng ",
"Ying ",
"Yeng ",
"Mong ",
"Jin ",
"Qi ",
"Pi ",
"Cey ",
"Ho ",
"Yu ",
"Zui ",
"Wo ",
"To ",
"Yin ",
"Yin ",
"Dui ",
"Ci ",
"Ho ",
"Jing ",
"Lam ",
"Cwun ",
"Ai ",
"Pok ",
"Thak ",
"Wei ",
"Pin ",
"Gu ",
"Cam ",
"Yeng ",
"Hin ",
"Kuo ",
"Fei ",
"",
"Boku ",
"Chen ",
"Wei ",
"Luo ",
"Zan ",
"Lye ",
"Li ",
],
"x70": [
"You ",
"Yang ",
"Lu ",
"Si ",
"Jie ",
"Hyeng ",
"Tok ",
"Wang ",
"Hui ",
"Sa ",
"Pan ",
"Sin ",
"Biao ",
"Chan ",
"Mo ",
"Lyu ",
"Jian ",
"Phok ",
"Sayk ",
"Cing ",
"Gu ",
"Pin ",
"Huo ",
"Xian ",
"Lo ",
"Qin ",
"Han ",
"Yeng ",
"Yung ",
"Lyek ",
"Ceng ",
"So ",
"Ying ",
"Sui ",
"Wei ",
"Hay ",
"Huai ",
"Hao ",
"Ce ",
"Lyong ",
"Loy ",
"Dui ",
"Fan ",
"Hu ",
"Loy ",
"",
"",
"Yeng ",
"Mi ",
"Ji ",
"Lyem ",
"Jian ",
"Ying ",
"Fen ",
"Lin ",
"Ik ",
"Chem ",
"Yue ",
"Chan ",
"Dai ",
"Yang ",
"Jian ",
"Lan ",
"Fan ",
"Shuang ",
"Yuan ",
"Zhuo ",
"Feng ",
"Sep ",
"Lei ",
"Lan ",
"Cong ",
"Qu ",
"Yong ",
"Qian ",
"Fa ",
"Kwan ",
"Que ",
"Yem ",
"Hao ",
"Hyeng ",
"Say ",
"Zan ",
"Lan ",
"Yan ",
"Li ",
"Mi ",
"Shan ",
"Than ",
"Dang ",
"Jiao ",
"Chan ",
"",
"Ho ",
"Pha ",
"Zhu ",
"Lan ",
"Lan ",
"Nang ",
"Man ",
"Luan ",
"Xun ",
"Xian ",
"Yem ",
"Gan ",
"Yem ",
"Yu ",
"Hwa ",
"Biao ",
"Mie ",
"Guang ",
"Ceng ",
"Hoy ",
"Xiao ",
"Xiao ",
"",
"Hong ",
"Lyeng ",
"Zao ",
"Zhuan ",
"Kwu ",
"Zha ",
"Xie ",
"Chi ",
"Cak ",
"Cay ",
"Cay ",
"Can ",
"Yang ",
"Ki ",
"Zhong ",
"Fen ",
"Niu ",
"Kyeng ",
"Mwun ",
"Po ",
"Yi ",
"Lo ",
"Chwi ",
"Pi ",
"Kai ",
"Pan ",
"Yem ",
"Kai ",
"Pang ",
"Mu ",
"Cho ",
"Liao ",
"Gui ",
"Hang ",
"Tun ",
"Guang ",
"Hun ",
"Ca ",
"Kwang ",
"Guang ",
"Wei ",
"Qiang ",
"",
"Da ",
"Xia ",
"Ceng ",
"Zhu ",
"Ke ",
"So ",
"Fu ",
"Ba ",
"Duo ",
"Duo ",
"Ling ",
"Zhuo ",
"Hyen ",
"Ke ",
"Than ",
"Pho ",
"Hyeng ",
"Pho ",
"Tai ",
"Tai ",
"Pyeng ",
"Yang ",
"Tong ",
"Han ",
"Cwu ",
"Cak ",
"Cem ",
"Wi ",
"Shi ",
"Lian ",
"Chi ",
"Huang ",
"",
"Hu ",
"Shuo ",
"Lan ",
"Jing ",
"Jiao ",
"Xu ",
"Xing ",
"Quan ",
"Lyel ",
"Hwan ",
"Yang ",
"Hyo ",
"Xiu ",
"Xian ",
"Yin ",
"O ",
"Zhou ",
"Yao ",
"Shi ",
"Kyey ",
"Tong ",
"Xue ",
"Zai ",
"Kai ",
"Hong ",
"Lak ",
"Xia ",
"Zhu ",
"Hwen ",
"Cung ",
"Po ",
"Yen ",
"Hui ",
"Guang ",
"Zhe ",
"Hoy ",
"Kao ",
"",
"Fan ",
"Shao ",
"Ye ",
"Hui ",
"",
"Tang ",
"Jin ",
"Re ",
"",
"Xi ",
"Fu ",
"Kyeng ",
"Che ",
"Pu ",
"Jing ",
"Zhuo ",
"Ting ",
"Wan ",
"Hai ",
"Phayng ",
"Lang ",
"Shan ",
"Hu ",
"Pong ",
"Chi ",
"Rong ",
],
"x71": [
"Hu ",
"",
"Shu ",
"He ",
"Hwun ",
"Ku ",
"Jue ",
"Xiao ",
"Xi ",
"En ",
"Han ",
"Zhuang ",
"Cwun ",
"Di ",
"Xie ",
"Ji ",
"Wu ",
"",
"Lua ",
"Han ",
"Yen ",
"Hwan ",
"Men ",
"Ju ",
"Chou ",
"Pay ",
"Pwun ",
"Lin ",
"Hon ",
"Hun ",
"Ton ",
"Xi ",
"Swi ",
"Mwu ",
"Hong ",
"Ju ",
"Fu ",
"Wo ",
"Cho ",
"Cong ",
"Feng ",
"Ping ",
"Qiong ",
"Ruo ",
"Xi ",
"Kyeng ",
"Hun ",
"Cak ",
"Yen ",
"Hyek ",
"Yi ",
"Jue ",
"Yu ",
"Gang ",
"Yen ",
"Pi ",
"Gu ",
"",
"Sheng ",
"Chang ",
"So ",
"",
"Nam ",
"Geng ",
"Wat ",
"Chen ",
"He ",
"Kui ",
"Zhong ",
"Duan ",
"Ha ",
"Hwi ",
"Feng ",
"Lyen ",
"Hwen ",
"Xing ",
"Hwang ",
"Jiao ",
"Cen ",
"Bi ",
"Yeng ",
"Ca ",
"Wi ",
"Tuan ",
"Tian ",
"Huy ",
"Nan ",
"Nan ",
"Chan ",
"Yen ",
"Jiong ",
"Jiong ",
"Wuk ",
"Mei ",
"Sal ",
"Wei ",
"Ye ",
"Xin ",
"Kyeng ",
"Rou ",
"May ",
"Hwan ",
"Hwu ",
"Co ",
"Oy ",
"Pen ",
"Qiu ",
"Sui ",
"Yang ",
"Lie ",
"Ca ",
"",
"Gao ",
"Gua ",
"Bao ",
"Hu ",
"Yun ",
"Xia ",
"",
"",
"Bian ",
"Gou ",
"Tui ",
"Tang ",
"Chao ",
"Sen ",
"En ",
"Bo ",
"Yep ",
"Xie ",
"Xi ",
"Wu ",
"Sik ",
"On ",
"He ",
"Hyo ",
"Huy ",
"Wun ",
"Wung ",
"Nai ",
"Shan ",
"",
"Yao ",
"Hwun ",
"Mi ",
"Lian ",
"Hyeng ",
"Wu ",
"Yong ",
"Kou ",
"En ",
"Qiang ",
"Liu ",
"Huy ",
"Bi ",
"Phyo ",
"Zong ",
"Lu ",
"Jian ",
"Swuk ",
"Sup ",
"Lou ",
"Pong ",
"Sui ",
"Ik ",
"Tong ",
"Jue ",
"Zong ",
"Wi ",
"Hu ",
"Yi ",
"Zhi ",
"O ",
"Wei ",
"Liao ",
"Sen ",
"Ou ",
"Yel ",
"Kyeng ",
"Man ",
"",
"Shang ",
"Cuan ",
"Zeng ",
"Jian ",
"Huy ",
"Huy ",
"Xi ",
"Yi ",
"Xiao ",
"Chi ",
"Huang ",
"Chan ",
"Yep ",
"Qian ",
"Yen ",
"Yan ",
"Xian ",
"Qiao ",
"Zun ",
"Tung ",
"Ton ",
"Shen ",
"Cho ",
"Fen ",
"Si ",
"Lyo ",
"Yu ",
"Lin ",
"Tong ",
"So ",
"Fen ",
"Pen ",
"Yen ",
"Sim ",
"Lan ",
"Mei ",
"Thang ",
"Yi ",
"Jing ",
"Men ",
"",
"",
"Yeng ",
"Wuk ",
"Yi ",
"Xue ",
"Lan ",
"Tai ",
"Co ",
"Chan ",
"Swu ",
"Xi ",
"Que ",
"Cong ",
"Lian ",
"Hwey ",
"Chok ",
"Sep ",
"Ling ",
"Wei ",
"Yi ",
"Xie ",
"Zhao ",
"Hui ",
"Tatsu ",
"Nung ",
"Lan ",
"Ru ",
"Huy ",
"Kao ",
"Hwun ",
"Sin ",
"Chou ",
"To ",
"Yo ",
],
"x72": [
"Hyek ",
"Lan ",
"Biao ",
"Rong ",
"Li ",
"Mo ",
"Phok ",
"Sel ",
"Lu ",
"La ",
"Ao ",
"Hwun ",
"Kwang ",
"Sak ",
"",
"Li ",
"Lo ",
"Jue ",
"Liao ",
"Yan ",
"Huy ",
"Xie ",
"Long ",
"Yep ",
"",
"Rang ",
"Yak ",
"Lan ",
"Cong ",
"Jue ",
"Tong ",
"Kwan ",
"",
"Che ",
"Mi ",
"Tang ",
"Lan ",
"Chok ",
"Laam ",
"Ling ",
"Chan ",
"Yu ",
"Co ",
"Lam ",
"Pha ",
"Cayng ",
"Pao ",
"Cheng ",
"Wen ",
"Ai ",
"Wi ",
"",
"Jue ",
"Cak ",
"Pwu ",
"Ye ",
"Pha ",
"Ta ",
"Ya ",
"Hyo ",
"Zu ",
"Sang ",
"I ",
"Qiang ",
"Sang ",
"Ge ",
"Cang ",
"Die ",
"Qiang ",
"Yong ",
"Cang ",
"Phyen ",
"Phan ",
"Pan ",
"Shao ",
"Cen ",
"Phay ",
"Du ",
"Chuang ",
"Tou ",
"Zha ",
"Bian ",
"Chep ",
"Pang ",
"Pak ",
"Chuang ",
"Yu ",
"",
"Tok ",
"A ",
"Thayng ",
"Wu ",
"Ushihen ",
"Pin ",
"Jiu ",
"Mo ",
"Tuo ",
"Mo ",
"Loy ",
"Ren ",
"Mang ",
"Fang ",
"Mao ",
"Mok ",
"Gang ",
"Mwul ",
"Yan ",
"Ge ",
"Bei ",
"Si ",
"Jian ",
"Ko ",
"You ",
"Ge ",
"Sayng ",
"Mu ",
"Ce ",
"Qian ",
"Quan ",
"Quan ",
"Ca ",
"Thuk ",
"Xi ",
"Mang ",
"Keng ",
"Kyen ",
"Wu ",
"Gu ",
"Se ",
"Li ",
"Li ",
"Pou ",
"Ji ",
"Gang ",
"Zhi ",
"Pwun ",
"Quan ",
"Run ",
"Du ",
"Ju ",
"Jia ",
"Ken ",
"Feng ",
"Pian ",
"Ke ",
"Ju ",
"Ho ",
"Chu ",
"Xi ",
"Bei ",
"Lak ",
"Jie ",
"Ma ",
"San ",
"Wei ",
"Li ",
"Ton ",
"Tong ",
"",
"Jiang ",
"Gi ",
"Li ",
"Tok ",
"Lie ",
"Pi ",
"Piao ",
"Bao ",
"Huy ",
"Chou ",
"Wei ",
"Kui ",
"Chou ",
"Kyen ",
"Hyun ",
"Ba ",
"Pem ",
"Qiu ",
"Ji ",
"Cai ",
"Cak ",
"An ",
"Hil ",
"Cang ",
"Guang ",
"Ma ",
"You ",
"Kang ",
"Bo ",
"Hou ",
"Ya ",
"Yin ",
"Huan ",
"Sang ",
"Yun ",
"Kwang ",
"Niu ",
"Cek ",
"Qing ",
"Zhong ",
"Mu ",
"Bei ",
"Pi ",
"Ju ",
"Ni ",
"Sheng ",
"Pao ",
"Ap ",
"Tuo ",
"Ho ",
"Ling ",
"Pi ",
"Pi ",
"Ni ",
"Ao ",
"You ",
"Kwu ",
"Yue ",
"Ce ",
"Dan ",
"Po ",
"Gu ",
"Xian ",
"Ning ",
"Huan ",
"Hen ",
"Kyo ",
"Hak ",
"Zhao ",
"Ji ",
"Swun ",
"Shan ",
"Ta ",
"Yung ",
"Swu ",
"Tong ",
"Lao ",
"Tok ",
"Xia ",
"Shi ",
"Hua ",
"Zheng ",
"Yu ",
"Sun ",
"Yu ",
"Phyey ",
"Mang ",
"Xi ",
"Kyen ",
"Li ",
"Hyep ",
"Un ",
"San ",
"Lang ",
"Phay ",
"Zhi ",
"Yan ",
],
"x73": [
"Sha ",
"Li ",
"Han ",
"Xian ",
"Jing ",
"Pai ",
"Fei ",
"Yao ",
"Ba ",
"Ki ",
"Yey ",
"Biao ",
"Yin ",
"Lai ",
"Xi ",
"Jian ",
"Qiang ",
"Kun ",
"Yan ",
"Kwa ",
"Zong ",
"Mi ",
"Chang ",
"Uy ",
"Cey ",
"Cayng ",
"Ya ",
"Mayng ",
"Si ",
"Col ",
"She ",
"Ryou ",
"Cen ",
"Luo ",
"Ho ",
"Zong ",
"Ji ",
"Oy ",
"Feng ",
"Wa ",
"Wen ",
"Seng ",
"Ce ",
"Myo ",
"Wei ",
"Yuan ",
"Hen ",
"Tuan ",
"Ya ",
"No ",
"Xie ",
"Jia ",
"Hwu ",
"Bian ",
"Yu ",
"Yu ",
"Mei ",
"Zha ",
"Yao ",
"Sun ",
"Bo ",
"Ming ",
"Hwal ",
"Wen ",
"Sou ",
"Ma ",
"Yuan ",
"Ay ",
"Ok ",
"Sa ",
"Hao ",
"",
"Yi ",
"Zhen ",
"Chuang ",
"Hao ",
"Man ",
"Jing ",
"Jiang ",
"Mu ",
"Cang ",
"Chan ",
"O ",
"Ao ",
"Hao ",
"Cui ",
"Fen ",
"Kwel ",
"Phyey ",
"Bi ",
"Huang ",
"Pu ",
"Lin ",
"Yu ",
"Tong ",
"Yao ",
"Lyo ",
"Shuo ",
"Xiao ",
"Swu ",
"Ton ",
"Xi ",
"Ge ",
"Juan ",
"Tok ",
"Yey ",
"Hoy ",
"Hem ",
"Hay ",
"Ta ",
"Xian ",
"Hwun ",
"Yeng ",
"Pin ",
"Hoyk ",
"Nou ",
"Meng ",
"Lyep ",
"Nao ",
"Kwang ",
"Swu ",
"Lu ",
"Tal ",
"Hen ",
"Mi ",
"Rang ",
"Huan ",
"Nao ",
"Luo ",
"Hem ",
"Qi ",
"Jue ",
"Hyen ",
"Miao ",
"Ca ",
"Sol ",
"Lo ",
"Ok ",
"Su ",
"Wang ",
"Qiu ",
"Ga ",
"Ceng ",
"Le ",
"Ba ",
"Ji ",
"Hong ",
"Di ",
"Chen ",
"Kan ",
"Kwu ",
"Wu ",
"Ki ",
"Ye ",
"Yang ",
"Ma ",
"Gong ",
"Wu ",
"Pwu ",
"Min ",
"Kay ",
"Ya ",
"Bin ",
"Bian ",
"Bang ",
"Yue ",
"Kyel ",
"Yun ",
"Kak ",
"Wan ",
"Jian ",
"May ",
"Dan ",
"Pin ",
"Wei ",
"Huan ",
"Xian ",
"Qiang ",
"Lyeng ",
"Tay ",
"Yi ",
"An ",
"Ping ",
"Cem ",
"Fu ",
"Hyeng ",
"Xi ",
"Pha ",
"Ci ",
"Gou ",
"Jia ",
"Shao ",
"Pak ",
"Ci ",
"Ka ",
"Ran ",
"Sheng ",
"Shen ",
"I ",
"Zu ",
"Ka ",
"Min ",
"San ",
"Liu ",
"Phil ",
"Cin ",
"Zhen ",
"Kak ",
"Fa ",
"Long ",
"Cin ",
"Jiao ",
"Jian ",
"Li ",
"Kwang ",
"Xian ",
"Zhou ",
"Kong ",
"Yan ",
"Xiu ",
"Yang ",
"Hwu ",
"Lak ",
"Su ",
"Cwu ",
"Qin ",
"Un ",
"Swun ",
"Po ",
"I ",
"Hyang ",
"Yao ",
"Xia ",
"Hyeng ",
"Kyu ",
"Chwung ",
"Xu ",
"Pan ",
"Phay ",
"",
"Dang ",
"Ei ",
"Hun ",
"Wen ",
"E ",
"Ceng ",
"Di ",
"Mwu ",
"Wu ",
"Seng ",
"Jun ",
"Mei ",
"Bei ",
"Ceng ",
"Hyen ",
"Chuo ",
],
"x74": [
"Han ",
"Sen ",
"Yan ",
"Kwu ",
"Hyen ",
"Lang ",
"Li ",
"Swu ",
"Fu ",
"Lyu ",
"Ye ",
"Xi ",
"Ling ",
"Li ",
"Cin ",
"Lian ",
"Suo ",
"Sa ",
"",
"Wan ",
"Dian ",
"Pin ",
"Zhan ",
"Cui ",
"Min ",
"Yu ",
"Ju ",
"Chim ",
"Lai ",
"Wen ",
"Sheng ",
"Wei ",
"Cen ",
"Swu ",
"Thak ",
"Pei ",
"Cayng ",
"Ho ",
"Ki ",
"E ",
"Kon ",
"Chang ",
"Ki ",
"Pong ",
"Wan ",
"Lu ",
"Cong ",
"Kwan ",
"Yem ",
"Co ",
"Pay ",
"Lim ",
"Kum ",
"Pi ",
"Pha ",
"Que ",
"Thak ",
"Qin ",
"Pep ",
"",
"Qiong ",
"Du ",
"Jie ",
"Hun ",
"Wu ",
"Mo ",
"Mei ",
"Chwun ",
"Sen ",
"Cey ",
"Seng ",
"Dai ",
"Rou ",
"Min ",
"Zhen ",
"Wi ",
"Ruan ",
"Huan ",
"Hay ",
"Chuan ",
"Jian ",
"Zhuan ",
"Yang ",
"Lian ",
"Quan ",
"Ha ",
"Duan ",
"Wen ",
"Ye ",
"No ",
"Ho ",
"Yeng ",
"Yu ",
"Hwang ",
"Se ",
"Sul ",
"Lyu ",
"",
"Yong ",
"Sway ",
"Yo ",
"On ",
"Wu ",
"Jin ",
"Cin ",
"Hyeng ",
"Ma ",
"Tao ",
"Liu ",
"Tang ",
"Li ",
"Lang ",
"Koy ",
"Cen ",
"Chang ",
"Cha ",
"Jue ",
"Zhao ",
"Yo ",
"Ai ",
"Bin ",
"Tu ",
"Chang ",
"Kun ",
"Zhuan ",
"Cong ",
"Kun ",
"Yi ",
"Choy ",
"Cong ",
"Ki ",
"Li ",
"Ying ",
"Suo ",
"Kwu ",
"Sen ",
"Ao ",
"Lyen ",
"Mwun ",
"Cang ",
"Yin ",
"",
"Ying ",
"Zhi ",
"Lu ",
"Wu ",
"Deng ",
"xiu ",
"Zeng ",
"Xun ",
"Qu ",
"Dang ",
"Lin ",
"Liao ",
"Qiong ",
"Su ",
"Hwang ",
"Koy ",
"Pak ",
"Kyeng ",
"Fan ",
"Cin ",
"Liu ",
"Ki ",
"",
"Kyeng ",
"Ai ",
"Pyek ",
"Chan ",
"Qu ",
"Co ",
"Tang ",
"Jiao ",
"Gun ",
"Tan ",
"Hui ",
"Hwan ",
"Se ",
"Swu ",
"Tian ",
"",
"Ye ",
"Jin ",
"Lu ",
"Bin ",
"Swu ",
"Wen ",
"Zui ",
"Lan ",
"Say ",
"Ji ",
"Sen ",
"Ruan ",
"Huo ",
"Gai ",
"Lei ",
"Du ",
"Li ",
"Cil ",
"Rou ",
"Li ",
"Zan ",
"Kyeng ",
"Zhe ",
"Gui ",
"Sui ",
"La ",
"Lyong ",
"Lu ",
"Li ",
"Zan ",
"Lan ",
"Yeng ",
"Mi ",
"Xiang ",
"Xi ",
"Kwan ",
"Dao ",
"Chan ",
"Hwan ",
"Kwa ",
"Bo ",
"Die ",
"Bao ",
"Ho ",
"Zhi ",
"Phyo ",
"Phan ",
"Rang ",
"Li ",
"Wa ",
"Dekaguramu ",
"Jiang ",
"Chen ",
"Fan ",
"Pen ",
"Fang ",
"Dan ",
"Ong ",
"Ou ",
"Deshiguramu ",
"Miriguramu ",
"Thon ",
"Hu ",
"Ling ",
"Yi ",
"Pyeng ",
"Ca ",
"Hekutogura ",
"Juan ",
"Chang ",
"Chi ",
"Sarake ",
"Dang ",
"Meng ",
"Pou ",
],
"x75": [
"Zhui ",
"Pyeng ",
"Bian ",
"Chwu ",
"Kyen ",
"Senchigura ",
"Ci ",
"Ying ",
"Qi ",
"Xian ",
"Lou ",
"Di ",
"Kwu ",
"Meng ",
"Cen ",
"Peng ",
"Lin ",
"Cung ",
"Wu ",
"Pyek ",
"Dan ",
"Ong ",
"Ying ",
"Yan ",
"Kam ",
"Dai ",
"Sim ",
"Chem ",
"Chem ",
"Han ",
"Sang ",
"Sayng ",
"Qing ",
"Shen ",
"San ",
"San ",
"Rui ",
"Sayng ",
"So ",
"Sen ",
"Yong ",
"Shuai ",
"Lu ",
"Po ",
"Yong ",
"Beng ",
"Feng ",
"Nyeng ",
"Cen ",
"Yu ",
"Kap ",
"Sin ",
"Zha ",
"Dian ",
"Fu ",
"Nam ",
"Cen ",
"Ping ",
"Ceng ",
"Hwa ",
"Ceng ",
"Kyen ",
"Zi ",
"Meng ",
"Pi ",
"Qi ",
"Liu ",
"Xun ",
"Liu ",
"Chang ",
"Mu ",
"Kyun ",
"Fan ",
"Fu ",
"Kyeng ",
"Cen ",
"Kyey ",
"Jie ",
"Kyen ",
"Oy ",
"Fu ",
"Cen ",
"Mu ",
"Tap ",
"Pan ",
"Jiang ",
"Wa ",
"Da ",
"Nan ",
"Lyu ",
"Pwun ",
"Cin ",
"Chwuk ",
"Mwu ",
"Mu ",
"Ce ",
"Cen ",
"Gai ",
"Phil ",
"Da ",
"Chi ",
"Lyak ",
"Hyu ",
"Lyak ",
"Pan ",
"Kesa ",
"Pen ",
"Hwa ",
"Yu ",
"Yu ",
"Mwu ",
"Cwun ",
"I ",
"Liu ",
"Yu ",
"Die ",
"Chou ",
"Hwa ",
"Tang ",
"Chuo ",
"Ki ",
"Wan ",
"Kang ",
"Sheng ",
"Chang ",
"Tuan ",
"Lei ",
"Ki ",
"Cha ",
"Liu ",
"Jou ",
"Tuan ",
"Lin ",
"Jiang ",
"Kang ",
"Cwu ",
"Pyek ",
"Die ",
"Chep ",
"Phil ",
"Nie ",
"Dan ",
"So ",
"So ",
"Zhi ",
"Uy ",
"Chuang ",
"Nai ",
"Ceng ",
"Bi ",
"Jie ",
"Liao ",
"Gang ",
"Hul ",
"Kwu ",
"Zhou ",
"Xia ",
"San ",
"Xu ",
"Nue ",
"Li ",
"Yang ",
"Chen ",
"Yu ",
"Pha ",
"Kay ",
"Jue ",
"Zhi ",
"Xia ",
"Chwi ",
"Bi ",
"Yek ",
"Li ",
"Zong ",
"Chuang ",
"Feng ",
"Zhu ",
"Pho ",
"Phi ",
"Kam ",
"A ",
"Ca ",
"Xie ",
"Qi ",
"Tal ",
"Cin ",
"Fa ",
"Zhi ",
"Tong ",
"Ce ",
"Cil ",
"Fei ",
"Kwu ",
"Dian ",
"Ka ",
"Hyen ",
"Zha ",
"Pyeng ",
"Ni ",
"Cung ",
"Yong ",
"Jing ",
"Cen ",
"Chong ",
"Tong ",
"I ",
"Hay ",
"Wei ",
"Hui ",
"Duo ",
"Yang ",
"Chi ",
"Chi ",
"Hun ",
"Ya ",
"Mei ",
"Twu ",
"Kyeng ",
"Xiao ",
"Thong ",
"Tu ",
"Mang ",
"Pi ",
"Xiao ",
"Suan ",
"Pu ",
"Li ",
"Ci ",
"Cwa ",
"Duo ",
"Wu ",
"Sa ",
"Lao ",
"Swu ",
"Huan ",
"Xian ",
"Yi ",
"Peng ",
"Zhang ",
"Guan ",
"Tam ",
"Fei ",
"Ma ",
"Lim ",
"Chi ",
"Ji ",
"Dian ",
"An ",
"Chi ",
"Pi ",
"Pi ",
"Min ",
"Ko ",
"Dui ",
"E ",
"Wi ",
],
"x76": [
"E ",
"Chwi ",
"Ya ",
"Zhu ",
"Cu ",
"Dan ",
"Shen ",
"Cong ",
"Kyey ",
"Yu ",
"Hou ",
"Phwung ",
"La ",
"Yang ",
"Shen ",
"Tu ",
"Yu ",
"Gua ",
"Wen ",
"Huan ",
"Ku ",
"Ha ",
"Yin ",
"Yi ",
"Lu ",
"So ",
"Jue ",
"Chi ",
"Xi ",
"Guan ",
"Yi ",
"On ",
"Chek ",
"Chang ",
"Pan ",
"Lei ",
"Lyu ",
"Chai ",
"Swu ",
"Hak ",
"Dian ",
"Da ",
"Pie ",
"Tan ",
"Zhang ",
"Biao ",
"Shen ",
"Cok ",
"La ",
"Yi ",
"Zong ",
"Chwu ",
"Cang ",
"Zhai ",
"Sou ",
"Suo ",
"Que ",
"Diao ",
"Lou ",
"Lwu ",
"Mo ",
"Jin ",
"Yin ",
"Ying ",
"Huang ",
"Fu ",
"Lyo ",
"Lyung ",
"Qiao ",
"Liu ",
"Lo ",
"Xian ",
"Phyey ",
"Tan ",
"Yin ",
"He ",
"Am ",
"Ban ",
"Kan ",
"Guan ",
"Guai ",
"Nong ",
"Yu ",
"Wei ",
"Yi ",
"Ong ",
"Pyek ",
"Lei ",
"Lye ",
"Shu ",
"Dan ",
"Lin ",
"Cen ",
"Lin ",
"Lai ",
"Bie ",
"Ji ",
"Chi ",
"Yang ",
"Xian ",
"Cel ",
"Cing ",
"Mak6",
"Lyek ",
"Kwak ",
"La ",
"Shaku ",
"Dian ",
"Sen ",
"Yeng ",
"Un ",
"Kwu ",
"Yong ",
"Than ",
"Cen ",
"Luo ",
"Luan ",
"Luan ",
"Bo ",
"",
"Kyey ",
"Po ",
"Pal ",
"Tung ",
"Pal ",
"Payk ",
"Payk ",
"Qie ",
"Kup ",
"Co ",
"Zao ",
"Mo ",
"Cek ",
"Pa ",
"Kay ",
"Hwang ",
"Gui ",
"Ci ",
"Ling ",
"Ko ",
"Mo ",
"Ji ",
"Kyo ",
"Peng ",
"Ko ",
"Ai ",
"E ",
"Ho ",
"Han ",
"Bi ",
"Hwan ",
"Chou ",
"Qian ",
"Xi ",
"Ay ",
"Hyo ",
"Ho ",
"Huang ",
"Ho ",
"Ze ",
"Cui ",
"Hao ",
"Xiao ",
"Ye ",
"Pha ",
"Hao ",
"Jiao ",
"Ai ",
"Xing ",
"Huang ",
"Li ",
"Piao ",
"Hak ",
"Jiao ",
"Phi ",
"Gan ",
"Pao ",
"Zhou ",
"Jun ",
"Qiu ",
"Cwun ",
"Que ",
"Sa ",
"Gu ",
"Kwun ",
"Jun ",
"Chwu ",
"Zha ",
"Gu ",
"Zhan ",
"Du ",
"Myeng ",
"Qi ",
"Ying ",
"Wu ",
"Pay ",
"Zhao ",
"Zhong ",
"Pwun ",
"He ",
"Yeng ",
"He ",
"Ik ",
"Bo ",
"Wan ",
"Hap ",
"Ang ",
"Zhan ",
"Yan ",
"Jian ",
"Hap ",
"Yu ",
"Hoy ",
"Fan ",
"Kay ",
"Dao ",
"Pan ",
"Po ",
"Qiu ",
"Seng ",
"To ",
"Lu ",
"Can ",
"Mayng ",
"Li ",
"Cin ",
"Xu ",
"Kam ",
"Pan ",
"Kwan ",
"An ",
"Lo ",
"Shu ",
"Zhou ",
"Thang ",
"An ",
"Ko ",
"Li ",
"Mok ",
"Ding ",
"Gan ",
"Yu ",
"Mayng ",
"Mang ",
"Cik ",
"Qi ",
"Ruan ",
"Tian ",
"Sang ",
"Swun ",
"Xin ",
"Hyey ",
"Pan ",
"Feng ",
"Swun ",
"Min ",
],
"x77": [
"Ming ",
"Seng ",
"Shi ",
"Yun ",
"Myen ",
"Pan ",
"Fang ",
"Myo ",
"Tham ",
"Mi ",
"Mo ",
"Kan ",
"Hyen ",
"Ou ",
"Shi ",
"Yang ",
"Zheng ",
"Yao ",
"Shen ",
"Huo ",
"Da ",
"Zhen ",
"Kuang ",
"Ju ",
"Shen ",
"Chi ",
"Sayng ",
"May ",
"Mo ",
"Zhu ",
"Cin ",
"Zhen ",
"Myen ",
"Di ",
"Yuan ",
"Die ",
"Yi ",
"Ca ",
"Ca ",
"Chao ",
"Cap ",
"Hyen ",
"Bing ",
"Mi ",
"Long ",
"Sui ",
"Dong ",
"Mi ",
"Die ",
"Yi ",
"Er ",
"Ming ",
"Swun ",
"Chi ",
"Kuang ",
"Kwen ",
"Mo ",
"Zhen ",
"Co ",
"Yang ",
"An ",
"Mo ",
"Cwung ",
"Mai ",
"Chak ",
"Zheng ",
"Mei ",
"Jun ",
"Shao ",
"Han ",
"Hwan ",
"Cey ",
"Cheng ",
"Cuo ",
"Juan ",
"E ",
"Wan ",
"Hyen ",
"Xi ",
"Kun ",
"Lai ",
"Jian ",
"Sem ",
"Tian ",
"Hun ",
"Wan ",
"Ling ",
"Shi ",
"Qiong ",
"Lie ",
"Ay ",
"Ceng ",
"Zheng ",
"Li ",
"Lai ",
"Swu ",
"Kwen ",
"Swu ",
"Swu ",
"Tok ",
"Bi ",
"Pi ",
"Mok ",
"Hun ",
"Yey ",
"Lu ",
"Yi ",
"Chep ",
"Chay ",
"Zhou ",
"Yu ",
"Hun ",
"Ma ",
"Xia ",
"Xing ",
"Xi ",
"Gun ",
"",
"Chun ",
"Ken ",
"Mei ",
"To ",
"Hou ",
"Xuan ",
"Ti ",
"Kyu ",
"Ko ",
"Yey ",
"Mao ",
"Xu ",
"Fa ",
"Wen ",
"Miao ",
"Chou ",
"Kui ",
"Mi ",
"Weng ",
"Kou ",
"Dang ",
"Cin ",
"Ke ",
"Swu ",
"Hal ",
"Qiong ",
"Mao ",
"Myeng ",
"Man ",
"Shui ",
"Ze ",
"Zhang ",
"Yey ",
"Diao ",
"Ou ",
"Mak ",
"Shun ",
"Cong ",
"Lou ",
"Chi ",
"Man ",
"Piao ",
"Tang ",
"Ji ",
"Mong ",
"",
"Run ",
"Pyel ",
"Xi ",
"Qiao ",
"Pu ",
"Zhu ",
"Cing ",
"Shen ",
"Swun ",
"Lyo ",
"Che ",
"Xian ",
"Kam ",
"Ye ",
"Xu ",
"Tong ",
"Mou ",
"Lin ",
"Kui ",
"Xian ",
"Ye ",
"Ay ",
"Hui ",
"Chem ",
"Kem ",
"Ko ",
"Zhao ",
"Kwu ",
"Wei ",
"Chou ",
"Sao ",
"Nyeng ",
"Xun ",
"Yao ",
"Huo ",
"Mong ",
"Mian ",
"Pin ",
"Mian ",
"Li ",
"Kuang ",
"Hwak ",
"Xuan ",
"Mian ",
"Huo ",
"Lu ",
"Meng ",
"Long ",
"Guan ",
"Man ",
"Xi ",
"Chok ",
"Tang ",
"Kem ",
"Chok ",
"Mo ",
"Kung ",
"Lin ",
"Yul ",
"Shuo ",
"Ce ",
"Hwak ",
"Si ",
"Uy ",
"Shen ",
"Ci ",
"Hou ",
"Sin ",
"Ying ",
"Kwu ",
"Zhou ",
"Jiao ",
"Cuo ",
"Tan ",
"Way ",
"Kyo ",
"Cung ",
"Huo ",
"Bai ",
"Sek ",
"Ding ",
"Qi ",
"Ji ",
"Zi ",
"Gan ",
"Wu ",
"Thak ",
"Kol ",
"Kang ",
"Sek ",
"Fan ",
"Kuang ",
],
"x78": [
"Dang ",
"Ma ",
"Sa ",
"Dan ",
"Jue ",
"Li ",
"Fu ",
"Min ",
"Nuo ",
"Huo ",
"Kang ",
"Zhi ",
"Chey ",
"Kan ",
"Jie ",
"Pwun ",
"E ",
"A ",
"Pi ",
"Zhe ",
"Yen ",
"Sway ",
"Zhuan ",
"Che ",
"Dun ",
"Pan ",
"Yan ",
"",
"Feng ",
"Fa ",
"Mo ",
"Zha ",
"Qu ",
"Yu ",
"La ",
"Tuo ",
"Tuo ",
"Ci ",
"Chay ",
"Chim ",
"Ai ",
"Fei ",
"Mu ",
"Zhu ",
"Lip ",
"Phyem ",
"Nu ",
"Ping ",
"Phayng ",
"Ling ",
"Pho ",
"Le ",
"Pha ",
"Bo ",
"Po ",
"Shen ",
"Za ",
"Nuo ",
"Lye ",
"Long ",
"Tong ",
"",
"Li ",
"Kou ",
"Chu ",
"Keng ",
"Quan ",
"Cwu ",
"Kuang ",
"Kyu ",
"E ",
"Nao ",
"Jia ",
"Lu ",
"Wei ",
"Ai ",
"Luo ",
"Ken ",
"Xing ",
"Yen ",
"Tong ",
"Peng ",
"Xi ",
"",
"Hong ",
"Shuo ",
"Xia ",
"Qiao ",
"",
"Wei ",
"Qiao ",
"",
"Kayng ",
"Cho ",
"Que ",
"Chan ",
"Lang ",
"Hong ",
"Yu ",
"Xiao ",
"Xia ",
"Mang ",
"Long ",
"Yong ",
"Cha ",
"Che ",
"E ",
"Lyu ",
"Kyeng ",
"Mang ",
"Hak ",
"Yen ",
"Sha ",
"Kun ",
"Yu ",
"",
"Kaki ",
"Lu ",
"Chen ",
"Jian ",
"Nue ",
"Song ",
"Zhuo ",
"Keng ",
"Pwung ",
"Yan ",
"Zhui ",
"Kong ",
"Ceng ",
"Ki ",
"Zong ",
"Qing ",
"Lin ",
"Jun ",
"Bo ",
"Ceng ",
"Min ",
"Diao ",
"Jian ",
"He ",
"Lok ",
"Ay ",
"Sway ",
"Cak ",
"Ling ",
"Pi ",
"Yin ",
"Tay ",
"Wu ",
"Qi ",
"Lun ",
"Wan ",
"Dian ",
"Gang ",
"Bei ",
"Qi ",
"Chen ",
"Ruan ",
"Yan ",
"Sel ",
"Ding ",
"Du ",
"Tuo ",
"Kal ",
"Yeng ",
"Bian ",
"Ke ",
"Pyek ",
"Oy ",
"Sek ",
"Zhen ",
"Duan ",
"Xia ",
"Thang ",
"Ti ",
"Nao ",
"Peng ",
"Jian ",
"Di ",
"Tan ",
"Cha ",
"Seki ",
"Qi ",
"",
"Feng ",
"Xuan ",
"Hwak ",
"Hwak ",
"Ma ",
"Gong ",
"Nyen ",
"Su ",
"E ",
"Ca ",
"Liu ",
"Si ",
"Tang ",
"Pang ",
"Hua ",
"Pi ",
"Oy ",
"Sang ",
"Loy ",
"Cha ",
"Zhen ",
"Xia ",
"Kyey ",
"Lyem ",
"Pan ",
"Ay ",
"Yun ",
"Dui ",
"Chayk ",
"Kay ",
"La ",
"",
"Qing ",
"Gun ",
"Cen ",
"Chan ",
"Qi ",
"Ao ",
"Peng ",
"Lu ",
"Nyo ",
"Kan ",
"Qiang ",
"Chen ",
"Yin ",
"Lei ",
"Biao ",
"Cek ",
"Ma ",
"Qi ",
"Choy ",
"Zong ",
"Kyeng ",
"Chuo ",
"",
"Ki ",
"Shan ",
"Lao ",
"Qu ",
"Zeng ",
"Tung ",
"Kan ",
"Xi ",
"Lin ",
"Ding ",
"Dian ",
"Kwang ",
"Pan ",
"Cap ",
"Kyo ",
"Di ",
"Li ",
],
"x79": [
"Kan ",
"Cho ",
"",
"Zhang ",
"Qiao ",
"Dun ",
"Xian ",
"Yu ",
"Zhui ",
"He ",
"Huo ",
"Zhai ",
"Loy ",
"Ke ",
"Cho ",
"Kup ",
"Que ",
"Tang ",
"Uy ",
"Jiang ",
"Pi ",
"Pi ",
"Ye ",
"Pin ",
"Qi ",
"Ay ",
"Kai ",
"Jian ",
"Yu ",
"Ruan ",
"Meng ",
"Pao ",
"Ci ",
"Bo ",
"",
"Mie ",
"Ca ",
"Xian ",
"Kwang ",
"Loy ",
"Lei ",
"Zhi ",
"Lye ",
"Lyek ",
"Pan ",
"Hwak ",
"Pao ",
"Ying ",
"Li ",
"Long ",
"Long ",
"Mo ",
"Bo ",
"Shuang ",
"Guan ",
"Lan ",
"Zan ",
"Yan ",
"Si ",
"Shi ",
"Lyey ",
"Reng ",
"Sa ",
"Yue ",
"Sa ",
"Ki ",
"Ta ",
"Ma ",
"Xie ",
"Yo ",
"Chen ",
"Ki ",
"Ki ",
"Ci ",
"Phayng ",
"Dui ",
"Zhong ",
"",
"Yi ",
"Shi ",
"Wu ",
"Zhi ",
"Tiao ",
"Pwul ",
"Pwu ",
"Pi ",
"Co ",
"Ci ",
"Suan ",
"Mei ",
"Co ",
"Ke ",
"Ho ",
"Chwuk ",
"Sin ",
"Swu ",
"Sa ",
"Chai ",
"Ni ",
"Lu ",
"Yu ",
"Sang ",
"Wu ",
"Co ",
"Phyo ",
"Zhu ",
"Gui ",
"Xia ",
"Zhi ",
"Cey ",
"Gao ",
"Zhen ",
"Gao ",
"Shui ",
"Jin ",
"Chen ",
"Gai ",
"Kun ",
"Di ",
"To ",
"Huo ",
"Tao ",
"Ki ",
"Gu ",
"Kwan ",
"Zui ",
"Ling ",
"Lok ",
"Phwum ",
"Kum ",
"Dao ",
"Zhi ",
"Lu ",
"Sen ",
"Bei ",
"Zhe ",
"Hui ",
"You ",
"Kyey ",
"In ",
"Zi ",
"Hwa ",
"Ceng ",
"Pok ",
"Yuan ",
"Wu ",
"Xian ",
"Yang ",
"Cey ",
"Yi ",
"Mei ",
"Si ",
"Di ",
"",
"Zhuo ",
"Zhen ",
"Yong ",
"Cik ",
"Gao ",
"Tang ",
"Si ",
"Ma ",
"Ta ",
"",
"Xuan ",
"Qi ",
"E ",
"Huy ",
"Ki ",
"Si ",
"Sen ",
"Tam ",
"Kuai ",
"Sui ",
"Lyey ",
"Nong ",
"Ni ",
"To ",
"Li ",
"Yang ",
"Yak ",
"Ti ",
"Zan ",
"Lei ",
"Rou ",
"Wu ",
"Ong ",
"Li ",
"Xie ",
"Kum ",
"Hwa ",
"Tok ",
"Swu ",
"Sa ",
"Ren ",
"Tok ",
"Zi ",
"Cha ",
"Kan ",
"Yi ",
"Xian ",
"Pyeng ",
"Nyen ",
"Chwu ",
"Qiu ",
"Chwung ",
"Fen ",
"Hao ",
"Yun ",
"Kwa ",
"Cho ",
"Zhi ",
"Geng ",
"Pi ",
"Ci ",
"Yu ",
"Pi ",
"Ku ",
"Ban ",
"Pi ",
"Ni ",
"Li ",
"You ",
"Co ",
"Pi ",
"Ba ",
"Lyeng ",
"Mal ",
"Ching ",
"Nian ",
"Cin ",
"Ang ",
"Zuo ",
"Cil ",
"Ci ",
"Chwul ",
"Ke ",
"Zi ",
"Huo ",
"Ji ",
"Ching ",
"Tong ",
"Zhi ",
"Huo ",
"He ",
"Yin ",
"Zi ",
"Zhi ",
"Kal ",
"Ren ",
"Du ",
"I ",
"Zhu ",
"Hui ",
"Nong ",
"Fu ",
],
"x7a": [
"Huy ",
"Ko ",
"Lang ",
"Fu ",
"Ze ",
"Sey ",
"Lu ",
"Kun ",
"Kan ",
"Geng ",
"Cey ",
"Ceng ",
"To ",
"Cho ",
"Sey ",
"Ya ",
"Lun ",
"Lu ",
"Gu ",
"Zuo ",
"Im ",
"Zhun ",
"Bang ",
"Phay ",
"Ji ",
"Cik ",
"Chi ",
"Kun ",
"Lung ",
"Peng ",
"Kwa ",
"Phwum ",
"Co ",
"Zu ",
"Yu ",
"Su ",
"Lue ",
"",
"Yi ",
"Sel ",
"Bian ",
"Ji ",
"Fu ",
"Bi ",
"Nuo ",
"Jie ",
"Cong ",
"Zong ",
"Xu ",
"Ching ",
"To ",
"On ",
"Lian ",
"Zi ",
"Wuk ",
"Cik ",
"Xu ",
"Cin ",
"Zhi ",
"To ",
"Ka ",
"Kyey ",
"Gao ",
"Ko ",
"Kok ",
"Rong ",
"Sui ",
"You ",
"Ji ",
"Kang ",
"Mok ",
"Shan ",
"Men ",
"Chi ",
"Ji ",
"Lu ",
"So ",
"Cek ",
"Yeng ",
"On ",
"Chwu ",
"Se ",
"",
"Yi ",
"Huang ",
"Qie ",
"Ji ",
"Swu ",
"Xiao ",
"Pu ",
"Jiao ",
"Zhuo ",
"Tong ",
"Sai ",
"Lu ",
"Sui ",
"Nong ",
"Sayk ",
"Yey ",
"Rang ",
"Nuo ",
"Yu ",
"",
"Ji ",
"Tui ",
"On ",
"Cheng ",
"Hwak ",
"Gong ",
"Lu ",
"Biao ",
"",
"Yang ",
"Zhuo ",
"Li ",
"Zan ",
"Hyel ",
"Al ",
"Kwu ",
"Qiong ",
"Xi ",
"Kwung ",
"Kong ",
"Yu ",
"Sen ",
"Ceng ",
"Yo ",
"Chen ",
"Twun ",
"Tol ",
"Lo ",
"Cel ",
"Chak ",
"Yo ",
"Phyem ",
"Bao ",
"Yo ",
"Bing ",
"Wa ",
"Zhu ",
"Jiao ",
"Qiao ",
"Diao ",
"Wu ",
"Gui ",
"Yao ",
"Cil ",
"Chang ",
"Yao ",
"Co ",
"Kyo ",
"Chang ",
"Kwun ",
"Xiao ",
"Cheng ",
"Kwu ",
"Cuan ",
"Wo ",
"Dan ",
"Kwul ",
"Kwa ",
"Zhui ",
"Xu ",
"Sol ",
"",
"Kui ",
"Dou ",
"",
"Yin ",
"Wa ",
"Wa ",
"Ya ",
"Yu ",
"Ju ",
"Kwung ",
"Yo ",
"Yao ",
"Tiao ",
"Chao ",
"Yu ",
"Tian ",
"Diao ",
"Kwu ",
"Liao ",
"Xi ",
"O ",
"Kyu ",
"Chang ",
"Zhao ",
"Fun ",
"Kwan ",
"Lyung ",
"Cheng ",
"Cui ",
"Piao ",
"Zao ",
"Chan ",
"Kyu ",
"Qiong ",
"Twu ",
"Co ",
"Long ",
"Cel ",
"Lip ",
"Chu ",
"Dekaritto ",
"Fou ",
"Kirorittoru ",
"Chu ",
"Hong ",
"Qi ",
"Miririttoru ",
"",
"Deshiritto ",
"Shu ",
"Myo ",
"Ju ",
"Cham ",
"Zhu ",
"Ling ",
"Lyong ",
"Pyeng ",
"Jing ",
"Kyeng ",
"Cang ",
"Hekutoritto ",
"Sa ",
"Cwun ",
"Hong ",
"Tong ",
"Song ",
"Jing ",
"Diao ",
"Yi ",
"Swu ",
"Jing ",
"Qu ",
"Kal ",
"Ping ",
"Tan ",
"Shao ",
"Zhuan ",
"Ceng ",
"Deng ",
"Cui ",
"Huai ",
"Kyeng ",
"Kan ",
"Kyeng ",
"Cwuk ",
"Chwuk ",
"Le ",
"Peng ",
"Wu ",
"Chi ",
"Kan ",
],
"x7b": [
"Mang ",
"Zhu ",
"Utsubo ",
"Du ",
"Ji ",
"Xiao ",
"Pha ",
"Suan ",
"Kup ",
"Zhen ",
"Co ",
"Swun ",
"A ",
"Zhui ",
"Yuan ",
"Hol ",
"Gang ",
"So ",
"Kum ",
"Pi ",
"Bi ",
"Jian ",
"Yi ",
"Dong ",
"Cem ",
"Sayng ",
"Xia ",
"Cek ",
"Zhu ",
"Na ",
"Thay ",
"Gu ",
"Lip ",
"Qie ",
"Min ",
"Bao ",
"Tiao ",
"Sa ",
"Pwu ",
"Chayk ",
"Pwun ",
"Pei ",
"Tal ",
"Zi ",
"Cey ",
"Lyeng ",
"Ze ",
"No ",
"Fu ",
"Gou ",
"Fan ",
"Ka ",
"Ge ",
"Pem ",
"Shi ",
"Mao ",
"Po ",
"Sey ",
"Jian ",
"Kong ",
"Long ",
"Souke ",
"Bian ",
"Luo ",
"Gui ",
"Qu ",
"Chi ",
"Yin ",
"Yao ",
"Sen ",
"Phil ",
"Qiong ",
"Kwal ",
"Tung ",
"Jiao ",
"Kun ",
"Cen ",
"Swun ",
"Ru ",
"Pel ",
"Kwang ",
"Chwuk ",
"Thong ",
"Kyey ",
"Tap ",
"Xing ",
"Chayk ",
"Zhong ",
"Kou ",
"Lai ",
"Bi ",
"Shai ",
"Dang ",
"Zheng ",
"Ce ",
"Pwu ",
"Kyun ",
"Tu ",
"Pa ",
"Li ",
"Lang ",
"Ke ",
"Kwan ",
"Kyen ",
"Han ",
"Thong ",
"Xia ",
"Zhi ",
"Seng ",
"Suan ",
"Se ",
"Zhu ",
"Zuo ",
"So ",
"Shao ",
"Ceng ",
"Ce ",
"Yen ",
"Gao ",
"Kuai ",
"Gan ",
"Chou ",
"Kyou ",
"Gang ",
"Yun ",
"O ",
"Qian ",
"So ",
"Jian ",
"Pu ",
"Lai ",
"Zou ",
"Bi ",
"Bi ",
"Bi ",
"Kay ",
"Chi ",
"Guai ",
"Yu ",
"Cen ",
"Zhao ",
"Ko ",
"Ho ",
"Cayng ",
"Jing ",
"Sha ",
"Chwu ",
"Lu ",
"Pak ",
"Ki ",
"Lin ",
"San ",
"Jun ",
"Pok ",
"Cha ",
"Gu ",
"Kong ",
"Kyem ",
"Quan ",
"Jun ",
"Chwu ",
"Kwan ",
"Yuan ",
"Ce ",
"Ju ",
"Bo ",
"Ze ",
"Qie ",
"Tuo ",
"Luo ",
"Tan ",
"Xiao ",
"Ruo ",
"Cen ",
"",
"Bian ",
"Sun ",
"Sang ",
"Xian ",
"Ping ",
"Cam ",
"Sheng ",
"Hu ",
"Shi ",
"Ce ",
"Yue ",
"Chun ",
"Lu ",
"Wu ",
"Dong ",
"So ",
"Ji ",
"Cel ",
"Hwang ",
"Xing ",
"Mei ",
"Pem ",
"Chui ",
"Cen ",
"Phyen ",
"Feng ",
"Chwuk ",
"Hong ",
"Hyep ",
"Hwu ",
"Qiu ",
"Miao ",
"Qian ",
"",
"Kui ",
"Sik ",
"Lou ",
"Wun ",
"He ",
"Tang ",
"Yue ",
"Chou ",
"Ko ",
"Fei ",
"Yak ",
"Zheng ",
"Kwu ",
"Nie ",
"Qian ",
"So ",
"Chan ",
"Gong ",
"Pang ",
"Tok ",
"Lyul ",
"Pi ",
"Zhuo ",
"Chu ",
"Sa ",
"Ci ",
"Zhu ",
"Qiang ",
"Long ",
"Lan ",
"Jian ",
"Bu ",
"Li ",
"Hui ",
"Phil ",
"Di ",
"Cong ",
"Yan ",
"Pong ",
"Sen ",
"Zhuan ",
"Pai ",
"Piao ",
"Dou ",
"Yu ",
"Myel ",
"Zhuan ",
],
"x7c": [
"Chayk ",
"Xi ",
"Guo ",
"Yi ",
"Hu ",
"Chan ",
"Kou ",
"Cok ",
"Ping ",
"Chwu ",
"Ji ",
"Kwey ",
"Su ",
"Lou ",
"Zha ",
"Lok ",
"Nian ",
"Sa ",
"Chan ",
"Sen ",
"Sa ",
"Le ",
"Duan ",
"Yana ",
"So ",
"Bo ",
"Mi ",
"Si ",
"Dang ",
"Liao ",
"Tan ",
"Cem ",
"Po ",
"Kan ",
"Min ",
"Kui ",
"Dai ",
"Jiao ",
"Deng ",
"Hwang ",
"Sun ",
"Lao ",
"Cam ",
"So ",
"Lu ",
"Shi ",
"Zan ",
"",
"Pai ",
"Hata ",
"Pai ",
"Gan ",
"Ju ",
"Du ",
"Lu ",
"Chem ",
"Pha ",
"Dang ",
"Sai ",
"Ke ",
"Long ",
"Chem ",
"Lyem ",
"Pwu ",
"Zhou ",
"Lai ",
"",
"Lam ",
"Kui ",
"Yu ",
"Yue ",
"Hao ",
"Zhen ",
"Tai ",
"Ti ",
"Mi ",
"Cwu ",
"Cek ",
"",
"Hata ",
"Tung ",
"Zhuan ",
"Cwu ",
"Fan ",
"Swu ",
"Zhou ",
"Sen ",
"Zhuo ",
"Teng ",
"Lu ",
"Lu ",
"Cen ",
"Thak ",
"Ying ",
"Yu ",
"Loy ",
"Long ",
"Shinshi ",
"Lian ",
"Lan ",
"Chem ",
"Yak ",
"Zhong ",
"Ke ",
"Lian ",
"Pyen ",
"Duan ",
"Zuan ",
"Li ",
"Si ",
"Luo ",
"Ying ",
"Yue ",
"Zhuo ",
"Yu ",
"Mi ",
"Di ",
"Fan ",
"Shen ",
"Zhe ",
"Shen ",
"Nu ",
"Xie ",
"Lyu ",
"Xian ",
"Zi ",
"In ",
"Cun ",
"",
"Chen ",
"Kume ",
"Pi ",
"Ban ",
"Wu ",
"Sha ",
"Kang ",
"Rou ",
"Pwun ",
"Bi ",
"Cui ",
"",
"Mo ",
"Chi ",
"Ta ",
"Ro ",
"Ba ",
"Lip ",
"Gan ",
"Ju ",
"Pak ",
"Mo ",
"Co ",
"Cem ",
"Zhou ",
"Li ",
"Su ",
"Tiao ",
"Li ",
"Xi ",
"Sok ",
"Hong ",
"Tong ",
"Ca ",
"Ce ",
"Wel ",
"Cwuk ",
"Lin ",
"Cang ",
"Payk ",
"",
"Fen ",
"Ji ",
"",
"Sukumo ",
"Lyang ",
"Xian ",
"Fu ",
"Lyang ",
"Chan ",
"Kayng ",
"Li ",
"Wel ",
"Lu ",
"Ju ",
"Qi ",
"Swu ",
"Bai ",
"Zhang ",
"Lin ",
"Cong ",
"Ceng ",
"Guo ",
"Kouji ",
"San ",
"San ",
"Tang ",
"Bian ",
"Yu ",
"Myen ",
"Hou ",
"Xu ",
"Zong ",
"Ho ",
"Jian ",
"Zan ",
"Ci ",
"Li ",
"Xie ",
"Fu ",
"Ni ",
"Bei ",
"Gu ",
"Xiu ",
"Ko ",
"Tang ",
"Kwu ",
"Sukumo ",
"Co ",
"Cang ",
"Tang ",
"Mi ",
"Sam ",
"Pwun ",
"Co ",
"Kang ",
"Jiang ",
"Mo ",
"San ",
"Can ",
"Na ",
"Xi ",
"Lyang ",
"Jiang ",
"Kuai ",
"Bo ",
"Huan ",
"",
"Zong ",
"Xian ",
"Na ",
"Tan ",
"Nie ",
"Lye ",
"Zuo ",
"Cek ",
"Nie ",
"Co ",
"Lan ",
"Sa ",
"Si ",
"Kyu ",
"Kyey ",
"Gong ",
"Zheng ",
"Kyu ",
"You ",
],
"x7d": [
"Ki ",
"Cha ",
"Cwu ",
"Swun ",
"Yak ",
"Hong ",
"Wu ",
"Hul ",
"Hwan ",
"Ren ",
"Mwun ",
"Mwun ",
"Qiu ",
"Nap ",
"Zi ",
"Tou ",
"Nyu ",
"Fou ",
"Jie ",
"Se ",
"Swun ",
"Pi ",
"Yin ",
"Sa ",
"Koyng ",
"Ci ",
"Kup ",
"Pwun ",
"Wun ",
"Ren ",
"Dan ",
"Jin ",
"So ",
"Pang ",
"Sayk ",
"Cui ",
"Jiu ",
"Zha ",
"Ha ",
"Jin ",
"Fu ",
"Zhi ",
"Ci ",
"Ca ",
"Cwu ",
"Hong ",
"Chal ",
"Lwu ",
"Sey ",
"Pwul ",
"Sel ",
"Sin ",
"Bei ",
"Ce ",
"Qu ",
"Ling ",
"Zhu ",
"So ",
"Kam ",
"Yang ",
"Fu ",
"Tuo ",
"Zhen ",
"Dai ",
"Chu ",
"Shi ",
"Cong ",
"Hyen ",
"Co ",
"Kyeng ",
"Pan ",
"Ju ",
"Mo ",
"Swul ",
"Zui ",
"Kou ",
"Kyeng ",
"Im ",
"Heng ",
"Xie ",
"Kyel ",
"Zhu ",
"Chou ",
"Gua ",
"Bai ",
"Cel ",
"Kwang ",
"Hu ",
"Ci ",
"Hwan ",
"Geng ",
"Co ",
"Hyel ",
"Ku ",
"Kyo ",
"Quan ",
"Gai ",
"Lak ",
"Hyen ",
"Pyeng ",
"Xian ",
"Fu ",
"Kup ",
"Tong ",
"Yung ",
"Co ",
"In ",
"Lei ",
"Xie ",
"Quan ",
"Se ",
"Gai ",
"Cil ",
"Thong ",
"Sa ",
"Kang ",
"Xiang ",
"Hui ",
"Cel ",
"Zhi ",
"Jian ",
"Kyen ",
"Chi ",
"Mian ",
"Zhen ",
"Lu ",
"Cheng ",
"Kwu ",
"Shu ",
"Bang ",
"Tong ",
"Cho ",
"Wan ",
"Qin ",
"Kyeng ",
"Xiu ",
"Ti ",
"Xiu ",
"Xie ",
"Hong ",
"Xi ",
"Fu ",
"Ceng ",
"Yu ",
"Dui ",
"Kun ",
"Fu ",
"Kyeng ",
"Hu ",
"Zhi ",
"Yan ",
"Jiong ",
"Pong ",
"Kyey ",
"Sok ",
"Kase ",
"Cong ",
"Lin ",
"Duo ",
"Li ",
"Lok ",
"Liang ",
"Cwu ",
"Kwen ",
"Shao ",
"Ki ",
"Ki ",
"Cwun ",
"Qi ",
"Wen ",
"Qian ",
"Sen ",
"Swu ",
"Yu ",
"Kyey ",
"Tao ",
"Kwan ",
"Kang ",
"Mang ",
"Pwung ",
"Chel ",
"Chay ",
"Guo ",
"Cui ",
"Lyun ",
"Liu ",
"Ki ",
"Than ",
"Bei ",
"Cak ",
"Lung ",
"Myen ",
"Qi ",
"Qie ",
"Tan ",
"Zong ",
"Kon ",
"Zou ",
"Yi ",
"Chi ",
"Xing ",
"Liang ",
"Kin ",
"Pi ",
"Yu ",
"Min ",
"Yu ",
"Chong ",
"Fan ",
"Lok ",
"Se ",
"Ying ",
"Zhang ",
"Kasuri ",
"Se ",
"Sang ",
"Ham ",
"Ke ",
"Sen ",
"Ruan ",
"Myen ",
"Cip ",
"Tan ",
"Zhong ",
"Chey ",
"Min ",
"Miao ",
"Yen ",
"Xie ",
"Bao ",
"Si ",
"Qiu ",
"Phyen ",
"Wan ",
"Geng ",
"Cong ",
"Myen ",
"Wei ",
"Fu ",
"Wi ",
"Yu ",
"Gou ",
"Myo ",
"Xie ",
"Lyen ",
"Zong ",
"Phyen ",
"Yun ",
"Yin ",
"Cey ",
"Gua ",
"Chi ",
"On ",
"Cheng ",
"Chan ",
"Dai ",
],
"x7e": [
"Xia ",
"Yen ",
"Chong ",
"Xu ",
"Jou ",
"Odoshi ",
"Geng ",
"Sen ",
"Yeng ",
"Cin ",
"Ayk ",
"Chwu ",
"Ni ",
"Bang ",
"Gu ",
"Pan ",
"Chwu ",
"Kyem ",
"Cuo ",
"Quan ",
"Shuang ",
"On ",
"Xia ",
"Choy ",
"Kyey ",
"Rong ",
"Tao ",
"Pak ",
"Yun ",
"Cin ",
"Ho ",
"Yo ",
"Kok ",
"Cay ",
"Tung ",
"Hyen ",
"Su ",
"Zhen ",
"Cong ",
"Tao ",
"Kou ",
"Cai ",
"Bi ",
"Pong ",
"Cu ",
"Li ",
"Chwuk ",
"Yen ",
"Xi ",
"Cong ",
"Lyu ",
"Kyen ",
"Kyen ",
"Man ",
"Zhi ",
"Lwu ",
"Mo ",
"Phyo ",
"Lian ",
"Mi ",
"Xuan ",
"Chong ",
"Cek ",
"Shan ",
"Sui ",
"Pen ",
"Shuai ",
"Pwung ",
"Yey ",
"So ",
"Mwu ",
"Yo ",
"Kang ",
"Hun ",
"Sem ",
"Kyey ",
"Zung ",
"Swu ",
"Ran ",
"Xuan ",
"Sey ",
"Qiao ",
"Cung ",
"Zuo ",
"Cik ",
"Sen ",
"San ",
"Lin ",
"Yu ",
"Pen ",
"Lyo ",
"Chuo ",
"Zun ",
"Jian ",
"Yo ",
"Chan ",
"Rui ",
"Swu ",
"Kwey ",
"Hua ",
"Chan ",
"Xi ",
"Kang ",
"Un ",
"Da ",
"Sung ",
"Hoy ",
"Kyey ",
"Se ",
"Kyen ",
"Jiang ",
"Hyen ",
"Co ",
"Cong ",
"Jie ",
"Jiao ",
"Bo ",
"Chan ",
"Ek ",
"Nao ",
"Sui ",
"Yek ",
"Shai ",
"Xu ",
"Kyey ",
"Pin ",
"Kyen ",
"Lan ",
"Pu ",
"Hwun ",
"Chan ",
"Qi ",
"Peng ",
"Li ",
"Mo ",
"Loy ",
"Hil ",
"Chan ",
"Kwang ",
"You ",
"Sok ",
"Lyu ",
"Xian ",
"Cen ",
"Kou ",
"Lu ",
"Chan ",
"Yeng ",
"Cay ",
"Xiang ",
"Sem ",
"Zui ",
"Chan ",
"Luo ",
"Xi ",
"To ",
"Lam ",
"Lei ",
"Lian ",
"Si ",
"Jiu ",
"Yu ",
"Hong ",
"Zhou ",
"Xian ",
"He ",
"Yue ",
"Ji ",
"Wan ",
"Kuang ",
"Ji ",
"Ren ",
"Wei ",
"Yun ",
"Koyng ",
"Chun ",
"Pi ",
"Sha ",
"Gang ",
"Na ",
"Ren ",
"Zong ",
"Lun ",
"Fen ",
"Zhi ",
"Wen ",
"Fang ",
"Ce ",
"Yin ",
"Niu ",
"Shu ",
"Xian ",
"Gan ",
"Xie ",
"Fu ",
"Lian ",
"Zu ",
"Shen ",
"Xi ",
"Zhi ",
"Zhong ",
"Zhou ",
"Ban ",
"Fu ",
"Zhuo ",
"Shao ",
"Yi ",
"Jing ",
"Dai ",
"Bang ",
"Rong ",
"Jie ",
"Ku ",
"Rao ",
"Die ",
"Heng ",
"Hui ",
"Gei ",
"Xuan ",
"Jiang ",
"Luo ",
"Jue ",
"Jiao ",
"Tong ",
"Geng ",
"Xiao ",
"Juan ",
"Xiu ",
"Xi ",
"Sui ",
"Tao ",
"Ji ",
"Ti ",
"Ji ",
"Xu ",
"Ling ",
"Ying ",
"Xu ",
"Qi ",
"Fei ",
"Chuo ",
"Zhang ",
"Gun ",
"Sheng ",
"Wei ",
"Mian ",
"Shou ",
"Beng ",
"Chou ",
"Tao ",
"Liu ",
"Quan ",
"Zong ",
"Zhan ",
"Wan ",
"Lu ",
],
"x7f": [
"Zhui ",
"Zi ",
"Ke ",
"Xiang ",
"Jian ",
"Mian ",
"Lan ",
"Ti ",
"Miao ",
"Qi ",
"On ",
"Hui ",
"Si ",
"Duo ",
"Duan ",
"Bian ",
"Sen ",
"Gou ",
"Zhui ",
"Huan ",
"Di ",
"Lu ",
"Bian ",
"Min ",
"Yuan ",
"Jin ",
"Fu ",
"Ru ",
"Zhen ",
"Feng ",
"Shuai ",
"Gao ",
"Chan ",
"Li ",
"Yi ",
"Jian ",
"Bin ",
"Piao ",
"Man ",
"Lei ",
"Ying ",
"Suo ",
"Mou ",
"Sao ",
"Xie ",
"Liao ",
"Shan ",
"Zeng ",
"Jiang ",
"Qian ",
"Zao ",
"Huan ",
"Jiao ",
"Zuan ",
"Pwu ",
"Sa ",
"Hang ",
"Fou ",
"Kyel ",
"Fou ",
"Ketsu ",
"Bo ",
"Pyeng ",
"Hang ",
"Diu6",
"Gang ",
"Ying ",
"Ayng ",
"Kyeng ",
"Ha ",
"Guan ",
"Cwun ",
"Tan ",
"Cang ",
"Qi ",
"Ong ",
"Ayng ",
"Loy ",
"Tam ",
"Lu ",
"Kwan ",
"Wang ",
"Wang ",
"Gang ",
"Mang ",
"Han ",
"",
"Luo ",
"Pwu ",
"Mi ",
"Fa ",
"Gu ",
"Zhu ",
"Ce ",
"Mao ",
"Ko ",
"Min ",
"Kang ",
"Ba ",
"Kway ",
"Ti ",
"Juan ",
"Pwu ",
"Lin ",
"Em ",
"Zhao ",
"Coy ",
"Kway ",
"Zhuo ",
"Yu ",
"Chi ",
"An ",
"Pel ",
"Nan ",
"Se ",
"Si ",
"Pi ",
"May ",
"Liu ",
"Pha ",
"Pel ",
"Li ",
"Chao ",
"Wei ",
"Phil ",
"Kyey ",
"Cung ",
"Tong ",
"Liu ",
"Ji ",
"Kyen ",
"Mi ",
"Zhao ",
"La ",
"Pi ",
"Ki ",
"Ki ",
"Luan ",
"Yang ",
"Mi ",
"Kang ",
"Ta ",
"Mi ",
"Yang ",
"You ",
"You ",
"Fen ",
"Ba ",
"Ko ",
"Yang ",
"Ko ",
"Kang ",
"Cang ",
"Gao ",
"Lyeng ",
"Yi ",
"Zhu ",
"Ce ",
"Swu ",
"Qian ",
"Yi ",
"Sen ",
"Rong ",
"Qun ",
"Kwun ",
"Qiang ",
"Huan ",
"Suo ",
"Sen ",
"Uy ",
"You ",
"Kang ",
"Xian ",
"Yu ",
"Geng ",
"Kal ",
"Tang ",
"Yuan ",
"Huy ",
"Fan ",
"Shan ",
"Fen ",
"Cen ",
"Lian ",
"Li ",
"Kayng ",
"Nou ",
"Qiang ",
"Chan ",
"Wu ",
"Gong ",
"Yi ",
"Chong ",
"Ong ",
"Fen ",
"Hong ",
"Chi ",
"Si ",
"Cui ",
"Fu ",
"Xia ",
"Pen ",
"Ik ",
"La ",
"Ik ",
"Pi ",
"Lyeng ",
"Liu ",
"Zhi ",
"Qu ",
"Sup ",
"Xie ",
"Sang ",
"Hup ",
"Hup ",
"Qi ",
"Qiao ",
"Hui ",
"Hwi ",
"So ",
"Se ",
"Hong ",
"Jiang ",
"Cek ",
"Chwi ",
"Pi ",
"Tao ",
"Sap ",
"Si ",
"Zhu ",
"Cen ",
"Xuan ",
"Shi ",
"Phyen ",
"Zong ",
"Wan ",
"Hwi ",
"Hou ",
"Hayk ",
"He ",
"Han ",
"Ko ",
"Piao ",
"Yey ",
"Lian ",
"Qu ",
"",
"Lin ",
"Pen ",
"Kyo ",
"Ko ",
"Pen ",
"Ik ",
"Hui ",
"Hyen ",
"Dao ",
],
"x80": [
"Yo ",
"Lo ",
"",
"Ko ",
"Mo ",
"Ca ",
"Ki ",
"Kwu ",
"Kwu ",
"Kwu ",
"Die ",
"Die ",
"I ",
"Shua ",
"Ruan ",
"Er ",
"Nay ",
"Tan ",
"Loy ",
"Ting ",
"Ca ",
"Kyeng ",
"Cho ",
"Mo ",
"Wun ",
"Pha ",
"Pi ",
"Chi ",
"Si ",
"Chu ",
"Ka ",
"Ju ",
"He ",
"Se ",
"Lao ",
"Lun ",
"Ji ",
"Tang ",
"Wu ",
"Lou ",
"Nwu ",
"Jiang ",
"Pang ",
"Ze ",
"Lwu ",
"Ki ",
"Lao ",
"Huo ",
"Wu ",
"Mo ",
"Huai ",
"I ",
"Zhe ",
"Ting ",
"Ya ",
"Da ",
"Song ",
"Qin ",
"Yun ",
"Chi ",
"Dan ",
"Tham ",
"Hong ",
"Kyeng ",
"Zhi ",
"",
"Nie ",
"Tam ",
"Zhen ",
"Che ",
"Lyeng ",
"Zheng ",
"You ",
"Wa ",
"Lyo ",
"Long ",
"Zhi ",
"Ning ",
"Tiao ",
"Er ",
"Ya ",
"Die ",
"Kwal ",
"",
"Lian ",
"Ho ",
"Seng ",
"Lie ",
"Ping ",
"Jing ",
"Chwi ",
"Bi ",
"Di ",
"Guo ",
"Mwun ",
"Xu ",
"Ping ",
"Cong ",
"Shikato ",
"",
"Ting ",
"Yu ",
"Cong ",
"Kui ",
"Ren ",
"Kui ",
"Cong ",
"Lyen ",
"Weng ",
"Kui ",
"Lyen ",
"Lyen ",
"Chong ",
"O ",
"Seng ",
"Yong ",
"Cheng ",
"Oy ",
"Sep ",
"Cik ",
"Dan ",
"Ning ",
"qie ",
"Ji ",
"Cheng ",
"Cheng ",
"Long ",
"Yul ",
"Yu ",
"Zhao ",
"Si ",
"Su ",
"I ",
"Swuk ",
"Sa ",
"Co ",
"Zhao ",
"Yuk ",
"Yi ",
"Luk ",
"Ki ",
"Qiu ",
"Ken ",
"Cao ",
"Ge ",
"Di ",
"Huan ",
"Hwang ",
"Yi ",
"Ren ",
"Cho ",
"Ru ",
"Cwu ",
"Yuan ",
"Twu ",
"Hang ",
"Rong ",
"Kan ",
"Cha ",
"Wo ",
"Chang ",
"Ko ",
"Ci ",
"Han ",
"Fu ",
"Pi ",
"Pwun ",
"Pei ",
"Pang ",
"Kyen ",
"Pang ",
"Swun ",
"You ",
"Nwul ",
"Hang ",
"Kung ",
"Ran ",
"Koyng ",
"Yuk ",
"Wen ",
"Hyo ",
"Ki ",
"Pi ",
"Qian ",
"Xi ",
"Xi ",
"Phyey ",
"Ken ",
"Jing ",
"Tai ",
"Shen ",
"Zhong ",
"Zhang ",
"Xie ",
"Shen ",
"Wi ",
"Cwu ",
"Die ",
"Tan ",
"Fei ",
"Ba ",
"Bo ",
"Kwu ",
"Tian ",
"Pay ",
"Gua ",
"Thay ",
"Zi ",
"Ku ",
"Zhi ",
"Ni ",
"Ping ",
"Zi ",
"Pwu ",
"Pan ",
"Zhen ",
"Xian ",
"Co ",
"Pay ",
"Kap ",
"Sheng ",
"Zhi ",
"Pho ",
"Mu ",
"Ke ",
"Ho ",
"Ke ",
"Yi ",
"Yun ",
"Se ",
"Yang ",
"Long ",
"Dong ",
"Ka ",
"Lu ",
"Jing ",
"Nu ",
"Yan ",
"Pang ",
"Kwa ",
"Yi ",
"Kwang ",
"Hai ",
"Ge ",
"Tong ",
"Zhi ",
"Xiao ",
"Hyung ",
"Hyung ",
"Er ",
"E ",
"Xing ",
"Pyen ",
"Nung ",
"Ca ",
"",
],
"x81": [
"Cheng ",
"Tiao ",
"Ci ",
"Cui ",
"Mei ",
"Hyep ",
"Chwi ",
"Hyep ",
"Mayk ",
"Mayk ",
"Chek ",
"Kyou ",
"Nin ",
"Kuai ",
"Sa ",
"Zang ",
"Qi ",
"Nao ",
"Mi ",
"Nong ",
"Luan ",
"Wan ",
"Pal ",
"Wen ",
"Wan ",
"Qiu ",
"Kak ",
"Kyeng ",
"Rou ",
"Heng ",
"Cuo ",
"Lie ",
"Shan ",
"Ting ",
"Mei ",
"Swun ",
"Sin ",
"Qian ",
"Te ",
"Choy ",
"Cu ",
"Swu ",
"Xin ",
"Thal ",
"Pao ",
"Cheng ",
"Nei ",
"Pho ",
"Twu ",
"Thal ",
"Niao ",
"Noy ",
"Pi ",
"Ko ",
"Gua ",
"Li ",
"Lian ",
"Chang ",
"Swu ",
"Jie ",
"Liang ",
"Zhou ",
"Pi ",
"Biao ",
"Lun ",
"Pyen ",
"Guo ",
"Kui ",
"Chui ",
"Dan ",
"Cen ",
"Nei ",
"Jing ",
"Jie ",
"Sek ",
"Ayk ",
"Yan ",
"Ren ",
"Sin ",
"Chuo ",
"Pwu ",
"Pwu ",
"Ke ",
"Pi ",
"Kang ",
"Wan ",
"Dong ",
"Pi ",
"Guo ",
"Zong ",
"Ding ",
"Wu ",
"Mei ",
"Ruan ",
"Zhuan ",
"Cil ",
"Cwu ",
"Gua ",
"Ou ",
"Di ",
"Am ",
"Seng ",
"Noy ",
"Yu ",
"Chuan ",
"Nan ",
"Yun ",
"Cong ",
"Rou ",
"Ak ",
"Sai ",
"Tu ",
"Yo ",
"Ken ",
"Wei ",
"Jiao ",
"Yu ",
"Jia ",
"Tan ",
"Phik ",
"Cang ",
"Pok ",
"Sen ",
"Ni ",
"Mian ",
"Wa ",
"Tung ",
"Thoy ",
"Pang ",
"Qian ",
"Lye ",
"Ol ",
"Swu ",
"Tang ",
"So ",
"Zhui ",
"Kyek ",
"Yi ",
"Pak ",
"Lyo ",
"Ji ",
"Pi ",
"Xie ",
"Ko ",
"Lu ",
"Bin ",
"",
"Chang ",
"Lu ",
"Guo ",
"Pang ",
"Chuai ",
"Phyo ",
"Jiang ",
"Pwu ",
"Tang ",
"Mak ",
"Sul ",
"Cen ",
"Lu ",
"Kyo ",
"Ying ",
"Lu ",
"Cil ",
"Tara ",
"Chun ",
"Lian ",
"Tong ",
"Phayng ",
"Ni ",
"Zha ",
"Liao ",
"Cui ",
"Gui ",
"Xiao ",
"Teng ",
"Pen ",
"Zhi ",
"Jiao ",
"Sen ",
"Mwu ",
"Chwey ",
"Run ",
"Xiang ",
"Swu ",
"Pwun ",
"Ung ",
"Dan ",
"Zhua ",
"Tam ",
"Hoy ",
"Nong ",
"Twun ",
"Lian ",
"Pi ",
"Yong ",
"Jue ",
"Chu ",
"Ek ",
"Juan ",
"La ",
"Lian ",
"Co ",
"Tun ",
"Gu ",
"Cey ",
"Cui ",
"Pin ",
"Xun ",
"No ",
"Huo ",
"Cang ",
"Xian ",
"Biao ",
"Xing ",
"Kuan ",
"Lap ",
"Yen ",
"Lye ",
"Huo ",
"Zang ",
"La ",
"Qu ",
"Cang ",
"Lyen ",
"Ni ",
"Zang ",
"Sin ",
"Qian ",
"Wa ",
"Guang ",
"Cang ",
"Lim ",
"Guang ",
"Ca ",
"Jiao ",
"El ",
"Chwi ",
"Ji ",
"Gao ",
"Chou ",
"Mian ",
"Nie ",
"Ci ",
"Chi ",
"Ge ",
"Chen ",
"Die ",
"Zhi ",
"Xiu ",
"Tay ",
"Cin ",
"Kwu ",
"Xian ",
"Yu ",
"Cha ",
],
"x82": [
"Yao ",
"Ye ",
"Yong ",
"Sek ",
"Sek ",
"Kwu ",
"Yu ",
"Ye ",
"Hung ",
"Ju ",
"Kwu ",
"Xin ",
"Sel ",
"Sa ",
"Sha ",
"Jiu ",
"Ci ",
"Tan ",
"Se ",
"Shi ",
"Tian ",
"Dan ",
"Pho ",
"Pu ",
"Kwan ",
"Hua ",
"Tan ",
"Chen ",
"Swun ",
"Xia ",
"Mwu ",
"Cwu ",
"Dao ",
"Kang ",
"Shan ",
"Yi ",
"",
"Pa ",
"Tai ",
"Fan ",
"Ban ",
"Chuan ",
"Hang ",
"Pang ",
"Pan ",
"Que ",
"Ro ",
"Zhong ",
"Jian ",
"Cang ",
"Ling ",
"Chwuk ",
"Ze ",
"Tha ",
"Pak ",
"Hyen ",
"Ka ",
"Sen ",
"Xia ",
"Lu ",
"Hong ",
"Pang ",
"Xi ",
"",
"Pwu ",
"Zao ",
"Feng ",
"Li ",
"Shao ",
"Ye ",
"Lang ",
"Ceng ",
"",
"Wei ",
"Bo ",
"Meng ",
"Nian ",
"Ke ",
"Huang ",
"Shou ",
"Zong ",
"Phyen ",
"Mao ",
"Die ",
"Dou6",
"Bang ",
"Cha ",
"Yi ",
"So ",
"Chang ",
"Co ",
"Lou ",
"Dai ",
"Sori ",
"Yao ",
"Tong ",
"Tou ",
"Dang ",
"Tan ",
"Lu ",
"Uy ",
"Jie ",
"Ham ",
"Huo ",
"Mong ",
"Qi ",
"Lo ",
"Lo ",
"Chan ",
"Shuang ",
"Kan ",
"Lyang ",
"Jian ",
"Kan ",
"Sayk ",
"Yan ",
"Pwul ",
"Ping ",
"Yem ",
"Yem ",
"Cho ",
"",
"Yi ",
"Le ",
"Ting ",
"Qiu ",
"Ay ",
"Nai ",
"Tiao ",
"Jiao ",
"Jie ",
"Peng ",
"Wan ",
"Yi ",
"Chai ",
"Mian ",
"Mi ",
"Gan ",
"Chen ",
"Wu ",
"Yu ",
"Cak ",
"Kwung ",
"Tu ",
"Xia ",
"Qi ",
"Mang ",
"Zi ",
"Hwey ",
"Sui ",
"Zhi ",
"Xiang ",
"Pi ",
"Pwu ",
"Twun ",
"Wei ",
"Wu ",
"Ci ",
"Qi ",
"Sam ",
"Wen ",
"Kem ",
"In ",
"Pwu ",
"Kou ",
"Kay ",
"Ho ",
"Se ",
"Kup ",
"Kum ",
"Ki ",
"Wen ",
"Pwun ",
"Pha ",
"Yey ",
"Sim ",
"Ki ",
"Hwa ",
"Hua ",
"Pang ",
"Wu ",
"Jue ",
"Gou ",
"Ci ",
"Wun ",
"Kun ",
"Ao ",
"Chwu ",
"Mo ",
"A ",
"Pi ",
"Ing ",
"Hang ",
"Cong ",
"Yin ",
"You ",
"Bian ",
"Yey ",
"Susa ",
"Wei ",
"Li ",
"Pi ",
"E ",
"Xian ",
"Chang ",
"Cang ",
"Meng ",
"Su ",
"Yi ",
"Wen ",
"Yem ",
"Lyeng ",
"Thay ",
"Cho ",
"Di ",
"Myo ",
"Qiong ",
"Li ",
"Yong ",
"Ka ",
"Mok ",
"Pei ",
"Pho ",
"Kwu ",
"Min ",
"I ",
"Yi ",
"Ke ",
"Pi ",
"Yak ",
"Ko ",
"Ce ",
"Ni ",
"Bo ",
"Bing ",
"Cem ",
"Qiu ",
"Yao ",
"Xian ",
"Pwun ",
"Hong ",
"Yeng ",
"Zha ",
"Tong ",
"Ce ",
"Die ",
"Nie ",
"Gan ",
"Hu ",
"Phyeng ",
"May ",
"Pwu ",
"Sheng ",
"Ko ",
"Phil ",
"Mi ",
],
"x83": [
"Pwul ",
"Cwul ",
"Mwu ",
"Pem ",
"Ka ",
"Mo ",
"Mao ",
"Ba ",
"Ca ",
"Mal ",
"Zi ",
"Di ",
"Chi ",
"Ji ",
"Jing ",
"Long ",
"",
"Niao ",
"",
"Xue ",
"Ying ",
"Qiong ",
"Ge ",
"Myeng ",
"Li ",
"Rong ",
"Yin ",
"Kan ",
"Chen ",
"Chay ",
"Chen ",
"Yu ",
"Xiu ",
"Zi ",
"Lie ",
"O ",
"Ji ",
"Gui ",
"Ce ",
"Chong ",
"Ca ",
"Gou ",
"Kwang ",
"Mang ",
"Si ",
"Jiao ",
"Jiao ",
"Pok ",
"Yu ",
"Swu ",
"Ca ",
"Kang ",
"Hoy ",
"In ",
"Ta ",
"Fa ",
"Yong ",
"Ye ",
"Chong ",
"Mang ",
"Tong ",
"Zhong ",
"",
"Zhu ",
"Swun ",
"Huan ",
"Kua ",
"Cen ",
"Gai ",
"Da ",
"Hyeng ",
"Hayng ",
"Chuan ",
"Cho ",
"Hyeng ",
"Er ",
"An ",
"Shou ",
"Chi ",
"Im ",
"Chen ",
"Cey ",
"Hwang ",
"Ping ",
"Li ",
"Jin ",
"Lao ",
"Shu ",
"Cang ",
"Da ",
"Jia ",
"Rao ",
"Bi ",
"Ze ",
"Qiao ",
"Hui ",
"Qi ",
"Dang ",
"",
"Rong ",
"Hun ",
"Ying ",
"Luo ",
"Ying ",
"Xun ",
"Jin ",
"Sun ",
"Yin ",
"Mai ",
"Hong ",
"Zhou ",
"Yao ",
"Du ",
"Wei ",
"Chu ",
"Twu ",
"Fu ",
"Ren ",
"Yin ",
"Ha ",
"Bi ",
"Bu ",
"Yun ",
"Cek ",
"Tu ",
"Sui ",
"Sui ",
"Cheng ",
"Chen ",
"Wu ",
"Bie ",
"Xi ",
"Geng ",
"Li ",
"Fu ",
"Zhu ",
"Mo ",
"Li ",
"Cang ",
"Ji ",
"Duo ",
"Qiu ",
"Sa ",
"Suo ",
"Chen ",
"Feng ",
"Ke ",
"May ",
"Meng ",
"Xing ",
"Kyeng ",
"Che ",
"Sin ",
"Jun ",
"Yen ",
"Ceng ",
"Diao ",
"Cwa ",
"Kwan ",
"Ham ",
"Yu ",
"Cuo ",
"Hyep ",
"Wang ",
"Chen ",
"Niu ",
"Shao ",
"Xian ",
"Lang ",
"Pwu ",
"A ",
"Mak ",
"Wen ",
"Jie ",
"Nan ",
"Mu ",
"Kan ",
"Lay ",
"Lian ",
"Shi ",
"Wo ",
"To ",
"Lian ",
"Huo ",
"You ",
"Ying ",
"Ying ",
"Nuc ",
"Chun ",
"Mang ",
"Mang ",
"Ca ",
"Wan ",
"Cheng ",
"Cek ",
"Qu ",
"Dong ",
"Kwan ",
"Zou ",
"Gu ",
"La ",
"Lok ",
"Kwuk ",
"Mi ",
"Kyun ",
"Nie ",
"Kun ",
"He ",
"Pu ",
"Chi ",
"Gao ",
"Kwa ",
"Pok ",
"Lun ",
"Chang ",
"Chou ",
"Swung ",
"Chui ",
"Zhan ",
"Men ",
"Chay ",
"Ba ",
"Li ",
"To ",
"Pha ",
"Ham ",
"Bao ",
"Qin ",
"Juan ",
"Xi ",
"Qin ",
"Di ",
"Jie ",
"Po ",
"Dang ",
"Kun ",
"Zhao ",
"Tai ",
"Geng ",
"Hwa ",
"Ko ",
"Lung ",
"Pi ",
"Jin ",
"Am ",
"Wang ",
"Pong ",
"Zhou ",
"Yan ",
"Ce ",
"Jian ",
"Lin ",
"Tan ",
"Swu ",
"Tian ",
"Dao ",
],
"x84": [
"Hu ",
"Qi ",
"He ",
"Chwey ",
"To ",
"Chwun ",
"Pi ",
"Cang ",
"Huan ",
"Fei ",
"Lay ",
"Che ",
"Mayng ",
"Phyeng ",
"Wi ",
"Dan ",
"Sha ",
"Huan ",
"Yan ",
"Yi ",
"Tiao ",
"Qi ",
"Wan ",
"Ce ",
"Nai ",
"Kutabireru ",
"Tuo ",
"Jiu ",
"Tie ",
"Luo ",
"",
"",
"Meng ",
"",
"Yaji ",
"",
"Ying ",
"Ying ",
"Ying ",
"Xiao ",
"Sa ",
"Chwu ",
"Ke ",
"Xiang ",
"Man ",
"Yu ",
"Yu ",
"Fu ",
"Lian ",
"Hwen ",
"Yuan ",
"Nan ",
"Ze ",
"Wa ",
"Chun ",
"Xiao ",
"Yu ",
"Phyen ",
"Mao ",
"An ",
"Ak ",
"Lak ",
"Ying ",
"Huo ",
"Gua ",
"Jiang ",
"Mian ",
"Zuo ",
"Zuo ",
"Ju ",
"Po ",
"Rou ",
"Xi ",
"Yep ",
"Am ",
"Qu ",
"Jian ",
"Fu ",
"Lu ",
"Jing ",
"Pen ",
"Phwung ",
"Hong ",
"Hong ",
"Hou ",
"Yan ",
"Tu ",
"Ce ",
"Zi ",
"Xiang ",
"Sim ",
"Kal ",
"Jie ",
"Jing ",
"Mi ",
"Huang ",
"Shen ",
"Pho ",
"Gai ",
"Tong ",
"Zhou ",
"Qian ",
"Wi ",
"Bo ",
"Wei ",
"Pha ",
"Kyey ",
"Ho ",
"Cang ",
"Ka ",
"Duan ",
"Yak ",
"Jun ",
"Chong ",
"Quan ",
"Wi ",
"Zhen ",
"Kyu ",
"Ting ",
"Hwun ",
"Xi ",
"Shi ",
"Cup ",
"Lan ",
"Zong ",
"Yao ",
"Yuan ",
"Mei ",
"Yun ",
"Shu ",
"Chey ",
"Zhuan ",
"Guan ",
"Sukumo ",
"Xue ",
"Chan ",
"Kai ",
"Kui ",
"",
"Cang ",
"Lou ",
"Wi ",
"Pai ",
"",
"Swu ",
"Yin ",
"Shi ",
"Chun ",
"Si ",
"Yun ",
"Zhen ",
"Lang ",
"Nu ",
"Mong ",
"He ",
"Que ",
"San ",
"Yuan ",
"Li ",
"Kwu ",
"Xi ",
"Pang ",
"Chu ",
"Xu ",
"Tu ",
"Liu ",
"Wo ",
"Zhen ",
"Chen ",
"Zu ",
"Po ",
"Cuo ",
"Yuan ",
"Chwu ",
"Yu ",
"Koy ",
"Pan ",
"Pu ",
"Pho ",
"Na ",
"Sak ",
"Xi ",
"Fen ",
"Yun ",
"Cung ",
"Kyem ",
"Cil ",
"Yak ",
"Chang ",
"En ",
"Mi ",
"Ho ",
"So ",
"Cin ",
"Myeng ",
"Huo ",
"Chwuk ",
"Liu ",
"Sek ",
"Gu ",
"Lang ",
"Yong ",
"Ong ",
"Kay ",
"Cuo ",
"Si ",
"Tang ",
"Luo ",
"Yok ",
"Sa ",
"Xian ",
"Pay ",
"Yao ",
"Gui ",
"Pi ",
"Zong ",
"Gun ",
"Za ",
"Swu ",
"Ce ",
"Hai ",
"Lan ",
"",
"Ji ",
"Li ",
"Can ",
"Lang ",
"Yu ",
"",
"Ying ",
"Mo ",
"Diao ",
"Tiao ",
"Mao ",
"Tong ",
"Zhu ",
"Pong ",
"Am ",
"Lyen ",
"Chong ",
"Xi ",
"Phyeng ",
"Qiu ",
"Jin ",
"Swun ",
"Jie ",
"Wei ",
"Tui ",
"Cao ",
"Yu ",
"Yi ",
"Ji ",
"Lyo ",
"Phil ",
"Lu ",
"Swuk ",
],
"x85": [
"Pwu ",
"Zhang ",
"Luo ",
"Jiang ",
"Man ",
"Yan ",
"Lung ",
"Ji ",
"Piao ",
"Gun ",
"Han ",
"Di ",
"Su ",
"Lu ",
"She ",
"Shang ",
"Di ",
"Myel ",
"Xun ",
"Man ",
"Pok ",
"Chey ",
"Cuo ",
"Ca ",
"Sam ",
"Xuan ",
"Wul ",
"Hu ",
"Ao ",
"Mi ",
"Lwu ",
"Cwu ",
"Zhong ",
"Chay ",
"Po ",
"Cang ",
"Mi ",
"Chong ",
"Co ",
"Hui ",
"Jun ",
"Yin ",
"Cem ",
"Yan ",
"So ",
"Um ",
"Kui ",
"Cin ",
"Hu ",
"Sha ",
"Kou ",
"Qian ",
"Ma ",
"Cang ",
"Sonoko ",
"Qiang ",
"Dou ",
"Lian ",
"Lin ",
"Kou ",
"Ai ",
"Phyey ",
"Li ",
"Wi ",
"Ji ",
"Tam ",
"Sheng ",
"Pen ",
"Meng ",
"Ou ",
"Chen ",
"Dian ",
"Xun ",
"Cho ",
"Yey ",
"Yey ",
"Lei ",
"Yu ",
"Kyo ",
"Chu ",
"Hua ",
"Jian ",
"Mai ",
"Wun ",
"Bao ",
"Yu ",
"Ke ",
"Lu ",
"Yo ",
"Hyey ",
"E ",
"Teng ",
"Fei ",
"Jue ",
"Chey ",
"Fa ",
"Ru ",
"Pwun ",
"Kui ",
"Swun ",
"Yu ",
"Ya ",
"Xu ",
"Fu ",
"Kwel ",
"Thang ",
"Mwu ",
"Tong ",
"Si ",
"So ",
"Xi ",
"Long ",
"On ",
"",
"Qi ",
"Jian ",
"Yun ",
"Sun ",
"Ling ",
"Yu ",
"Xia ",
"Yong ",
"Cup ",
"Hong ",
"Si ",
"Nong ",
"Loy ",
"Xuan ",
"On ",
"Yu ",
"Xi ",
"Hao ",
"Pak ",
"Hao ",
"Ay ",
"Mi ",
"Hui ",
"Yey ",
"Kyey ",
"Ci ",
"Hyang ",
"Luan ",
"Mie ",
"Uy ",
"Leng ",
"Kang ",
"Can ",
"Shen ",
"Cang ",
"Lian ",
"Ke ",
"Wen ",
"Da ",
"Chi ",
"Tang ",
"Sel ",
"Pyek ",
"Tam ",
"Sun ",
"Lian ",
"Fan ",
"Ding ",
"Jie ",
"Gu ",
"Hay ",
"Chok ",
"Chen ",
"Kao ",
"Hwung ",
"Sal ",
"Sin ",
"Hwun ",
"Yak ",
"Hai ",
"Sou ",
"Se ",
"Hwun ",
"Dui ",
"Pin ",
"Wei ",
"Neng ",
"Chou ",
"Mai ",
"Ru ",
"Piao ",
"Tai ",
"Cey ",
"Zao ",
"Chen ",
"Zhen ",
"Er ",
"Ni ",
"Ying ",
"Ko ",
"Cong ",
"Xiao ",
"Qi ",
"Fa ",
"Jian ",
"Xu ",
"Kui ",
"Ca ",
"Bian ",
"Diao ",
"Mi ",
"Lam ",
"Sin ",
"Cang ",
"Myo ",
"Qiong ",
"Qie ",
"Xian ",
"",
"Wu ",
"Xian ",
"Su ",
"Lu ",
"Yi ",
"Xu ",
"Xie ",
"Lye ",
"Yey ",
"La ",
"Lei ",
"Xiao ",
"Di ",
"Zhi ",
"Bei ",
"Tung ",
"Yak ",
"Mo ",
"Huan ",
"Biao ",
"Pen ",
"Swu ",
"Tan ",
"Tui ",
"Qiong ",
"Qiao ",
"Wei ",
"Liu ",
"Hui ",
"",
"Gao ",
"On ",
"",
"Li ",
"Ce ",
"Chu ",
"Ay ",
"Lin ",
"Co ",
"Xuan ",
"Chin ",
"Lai ",
"Kwak ",
],
"x86": [
"Thak ",
"Wu ",
"Yey ",
"Rui ",
"Ki ",
"Heng ",
"Lo ",
"So ",
"Tui ",
"Mang ",
"On ",
"Pin ",
"Yu ",
"Xun ",
"Ji ",
"Jiong ",
"Xian ",
"Mo ",
"Hagi ",
"Su ",
"Jiong ",
"",
"El ",
"Pyek ",
"Yang ",
"Yi ",
"Sen ",
"Yu ",
"Ju ",
"Lian ",
"Lian ",
"Yin ",
"Qiang ",
"Ying ",
"Long ",
"Tong ",
"Wei ",
"Yue ",
"Ling ",
"Ke ",
"Yao ",
"Pen ",
"Mi ",
"Lan ",
"Kui ",
"Lan ",
"Ji ",
"Thang ",
"Katsura ",
"Lei ",
"Lei ",
"Hua ",
"Feng ",
"Zhi ",
"Wei ",
"Kui ",
"Zhan ",
"Huai ",
"Li ",
"Ji ",
"Mi ",
"Lei ",
"Huai ",
"La ",
"Ji ",
"Ki ",
"Lu ",
"Jian ",
"San ",
"",
"Lei ",
"Quan ",
"Xiao ",
"Yi ",
"Luan ",
"Men ",
"Bie ",
"Hu ",
"Ho ",
"Lu ",
"Hak ",
"Lu ",
"Si ",
"Hyo ",
"Ken ",
"Che ",
"Hu ",
"Xu ",
"Cuo ",
"Fu ",
"He ",
"He ",
"Lo ",
"Hu ",
"Wu ",
"Ho ",
"Jiao ",
"Ju ",
"Guo ",
"Bao ",
"Yan ",
"Zhan ",
"Zhan ",
"Hyu ",
"Ban ",
"Xi ",
"Shu ",
"Hwey ",
"Kyu ",
"Diao ",
"Ji ",
"Kyu ",
"Cheng ",
"Sul ",
"",
"Di ",
"Zhe ",
"She ",
"Yu ",
"Gan ",
"Ca ",
"Hong ",
"Hwey ",
"Meng ",
"Ge ",
"Sui ",
"Xia ",
"Chai ",
"Shi ",
"Yi ",
"Ma ",
"Xiang ",
"Fang ",
"E ",
"Pa ",
"Chi ",
"Qian ",
"Wen ",
"Mwun ",
"Yey ",
"Pang ",
"Pi ",
"Yue ",
"Yue ",
"Jun ",
"Qi ",
"Tong ",
"In ",
"Qi ",
"Chen ",
"Yuan ",
"Jue ",
"Hui ",
"Qin ",
"Qi ",
"Zhong ",
"Ya ",
"Ci ",
"Mu ",
"Wang ",
"Fen ",
"Fen ",
"Hang ",
"Kong ",
"Co ",
"Fu ",
"Ran ",
"Jie ",
"Pwu ",
"Chi ",
"Twu ",
"Piao ",
"Xian ",
"Ni ",
"Te ",
"Kwu ",
"Yu ",
"Chayk ",
"Ping ",
"Chi ",
"Yu ",
"He ",
"Han ",
"Ju ",
"Li ",
"Fu ",
"Ran ",
"Zha ",
"Gou ",
"Pi ",
"Bo ",
"Xian ",
"Cwu ",
"Diao ",
"Bie ",
"Bing ",
"Gu ",
"Cem ",
"Ce ",
"Sa ",
"Tie ",
"Lyeng ",
"Gu ",
"Tan ",
"Gu ",
"Hyeng ",
"Li ",
"Cheng ",
"Qu ",
"Mou ",
"Ge ",
"Ci ",
"Hoy ",
"Hui ",
"Mang ",
"Fu ",
"Yang ",
"Wa ",
"Lie ",
"Cwu ",
"Yi ",
"Xian ",
"Hwal ",
"Kyo ",
"Li ",
"Yi ",
"Ping ",
"Kil ",
"Hap ",
"She ",
"Yi ",
"Wang ",
"Mo ",
"Kong ",
"Qie ",
"Gui ",
"Kong ",
"Cil ",
"Man ",
"Ebi ",
"Zhi ",
"Jia ",
"Rao ",
"Si ",
"Qi ",
"Xing ",
"Lie ",
"Qiu ",
"So ",
"Yong ",
"Jia ",
"Sey ",
"Che ",
"Bai ",
"A ",
"Han ",
],
"x87": [
"Chok ",
"Xuan ",
"Pong ",
"Sin ",
"Zhen ",
"Fu ",
"Hyen ",
"Zhe ",
"O ",
"Pwu ",
"Li ",
"Lang ",
"Bi ",
"Chu ",
"Yuan ",
"You ",
"Jie ",
"Tan ",
"Yan ",
"Ting ",
"Dian ",
"Sey ",
"Hoy ",
"Gua ",
"Ci ",
"Song ",
"Pi ",
"Ju ",
"Mil ",
"Ki ",
"Qi ",
"Yu ",
"Jun ",
"Sa ",
"Meng ",
"Qiang ",
"Si ",
"Sek ",
"Lun ",
"Li ",
"Cep ",
"Co ",
"Tao ",
"Kun ",
"Gan ",
"Han ",
"Yu ",
"Bang ",
"Fei ",
"Pi ",
"Wei ",
"Dun ",
"Chek ",
"Yen ",
"Su ",
"Kwen ",
"Qian ",
"Rui ",
"Yey ",
"Cheng ",
"Wei ",
"Liang ",
"Guo ",
"Wen ",
"Tong ",
"E ",
"Ban ",
"Di ",
"Wang ",
"Can ",
"Yang ",
"Ying ",
"Guo ",
"Sen ",
"",
"Lap ",
"Kwa ",
"Ji ",
"Kal ",
"Ting ",
"Mai ",
"Xu ",
"Mian ",
"Yu ",
"Jie ",
"Sik ",
"Xuan ",
"Hwang ",
"Yan ",
"Phyen ",
"Rou ",
"Wei ",
"Fu ",
"Yuan ",
"Mei ",
"Wi ",
"Pok ",
"Ruan ",
"Xie ",
"You ",
"Yu ",
"Mo ",
"Ha ",
"Ying ",
"Sul ",
"Chong ",
"Tang ",
"Zhu ",
"Zong ",
"Ti ",
"Pok ",
"Yuan ",
"Hui ",
"Meng ",
"La ",
"Du ",
"Hwu ",
"Qiu ",
"Cep ",
"Li ",
"Wa ",
"Yun ",
"Ju ",
"Nan ",
"Lou ",
"Qun ",
"Rong ",
"Ying ",
"Jiang ",
"",
"Lang ",
"Pang ",
"Si ",
"Xi ",
"Ci ",
"Xi ",
"Yuan ",
"Weng ",
"Lian ",
"Sou ",
"Pan ",
"Yung ",
"Rong ",
"Ji ",
"Wu ",
"Qiu ",
"Han ",
"Qin ",
"Yi ",
"Bi ",
"Hua ",
"Tang ",
"Uy ",
"Du ",
"Nai ",
"He ",
"Hu ",
"Hui ",
"Ma ",
"Myeng ",
"Yi ",
"Wen ",
"Hyeng ",
"Tung ",
"Yu ",
"Cang ",
"So ",
"Ebi ",
"Man ",
"",
"Shang ",
"Shi ",
"Cao ",
"Li ",
"Di ",
"Ao ",
"Lu ",
"Wei ",
"Zhi ",
"Tang ",
"Cin ",
"Piao ",
"Qu ",
"Pi ",
"Yu ",
"Jian ",
"La ",
"Lwu ",
"Qin ",
"Cong ",
"Yin ",
"Jiang ",
"Sol ",
"Wen ",
"Jiao ",
"Wan ",
"Chip ",
"Zhe ",
"Ma ",
"Ma ",
"Guo ",
"Liu ",
"Mo ",
"Sil ",
"Cong ",
"Li ",
"Man ",
"Xiao ",
"Tou ",
"Zhang ",
"Mang ",
"Xiang ",
"Mo ",
"Zui ",
"Si ",
"Qiu ",
"Te ",
"Zhi ",
"Phayng ",
"Phayng ",
"Jiao ",
"Qu ",
"Bie ",
"Liao ",
"Pan ",
"Gui ",
"Xi ",
"Ki ",
"Zhuan ",
"Huang ",
"Fei ",
"Lao ",
"Jue ",
"Jue ",
"Hui ",
"Yin ",
"Sen ",
"Jiao ",
"Shan ",
"Yo ",
"Xiao ",
"Mou ",
"Chwung ",
"Xun ",
"Si ",
"",
"Cheng ",
"Tang ",
"Li ",
"Hay ",
"Shan ",
"Uy ",
"Jing ",
"Da ",
"Sem ",
"Qi ",
],
"x88": [
"Ci ",
"Xiang ",
"She ",
"La ",
"Qin ",
"Sung ",
"Chai ",
"Li ",
"Ze ",
"Xuan ",
"Lian ",
"Zhu ",
"Ze ",
"Kal ",
"Mang ",
"Xie ",
"Qi ",
"Yeng ",
"Jian ",
"Meng ",
"Hao ",
"Yu ",
"Huo ",
"Zhuo ",
"Jie ",
"Bin ",
"He ",
"Mie ",
"Fan ",
"Lei ",
"Jie ",
"Lap ",
"Mi ",
"Lye ",
"Cwun ",
"Lye ",
"Qiu ",
"Nie ",
"Lu ",
"Du ",
"Xiao ",
"Zhu ",
"Long ",
"Li ",
"Long ",
"Feng ",
"Ye ",
"Beng ",
"Shang ",
"Ko ",
"Kyen ",
"Ying ",
"",
"Xi ",
"Cam ",
"Qu ",
"Quan ",
"Twu ",
"Can ",
"Man ",
"Jue ",
"Jie ",
"Zhu ",
"Zha ",
"Hyel ",
"Huang ",
"Niu ",
"Pei ",
"Nyuk ",
"Hun ",
"Cwung ",
"Mo ",
"Er ",
"Ke ",
"Myel ",
"Xi ",
"Hayng ",
"Yen ",
"Kan ",
"Yuan ",
"",
"Ling ",
"Hyen ",
"Swul ",
"Xian ",
"Tong ",
"Long ",
"Ka ",
"Xian ",
"A ",
"Hu ",
"Wi ",
"Dao ",
"Chwung ",
"Wi ",
"Dao ",
"Zhun ",
"Hyeng ",
"Kwu ",
"Uy ",
"Koromohen ",
"Bu ",
"Gan ",
"Yu ",
"Phyo ",
"Cha ",
"Yi ",
"Sam ",
"Chen ",
"Fu ",
"Kon ",
"Fen ",
"Soy ",
"Jie ",
"Nap ",
"Zhong ",
"Dan ",
"Ri ",
"Zhong ",
"Chwung ",
"Xie ",
"Qi ",
"Xie ",
"Ran ",
"Zhi ",
"Im ",
"Kum ",
"Kum ",
"Jun ",
"Wen ",
"Myey ",
"Chai ",
"Ao ",
"Niao ",
"Wi ",
"Ran ",
"Ka ",
"Tuo ",
"Ling ",
"Tay ",
"Bao ",
"Pho ",
"Yao ",
"Zuo ",
"Bi ",
"Shao ",
"Tan ",
"Ju ",
"He ",
"Shu ",
"Swu ",
"Cin ",
"Yi ",
"Pa ",
"Bo ",
"Ce ",
"Wa ",
"Pwu ",
"Kon ",
"Zhi ",
"Zhi ",
"Ran ",
"Pen ",
"Yi ",
"Mwu ",
"",
"Na ",
"Kou ",
"Xian ",
"Chan ",
"Ke ",
"Phi ",
"Gun ",
"Xi ",
"Ne ",
"Bo ",
"Horo ",
"Pok ",
"Yi ",
"Chi ",
"Ko ",
"Ren ",
"Jiang ",
"Jia ",
"Cun ",
"Mo ",
"Jie ",
"Er ",
"Luo ",
"Ru ",
"Zhu ",
"Gui ",
"In ",
"Cay ",
"Lyel ",
"Kamishimo ",
"Yuki ",
"Cang ",
"Dang ",
"Seot ",
"Kun ",
"Ken ",
"Nyo ",
"Shu ",
"Jia ",
"Kon ",
"Cheng ",
"Li ",
"Juan ",
"Shen ",
"Pwu ",
"Ge ",
"Yey ",
"Yu ",
"Zhen ",
"Liu ",
"Kwu ",
"Kwun ",
"Ji ",
"Yi ",
"Po ",
"Cang ",
"Shui ",
"Sa ",
"Qun ",
"Li ",
"Lian ",
"Lian ",
"Ku ",
"Jian ",
"Fou ",
"Chan ",
"Pi ",
"Gun ",
"Tao ",
"Yuan ",
"Ling ",
"Chi ",
"Chang ",
"Cwu ",
"Duo ",
"Phyo ",
"Liang ",
"Sang ",
"Pay ",
"Pey ",
"Fei ",
"Yuan ",
"La ",
"Kwa ",
"Yan ",
"Du ",
"Xi ",
"Cey ",
"Ke ",
"Qi ",
],
"x89": [
"Ji ",
"Zhi ",
"Gua ",
"Ken ",
"Che ",
"Ti ",
"Ti ",
"Pok ",
"Chong ",
"Xie ",
"Phyen ",
"Die ",
"Kon ",
"Duan ",
"Xiu ",
"Xiu ",
"Kal ",
"Yuan ",
"Pho ",
"Po ",
"Fu ",
"Yu ",
"Tuan ",
"Yan ",
"Wi ",
"Pey ",
"Ce ",
"Lu ",
"Ena ",
"Tan ",
"Yun ",
"Da ",
"Gou ",
"Da ",
"Huai ",
"Rong ",
"Yuan ",
"Yo ",
"Nai ",
"Kyeng ",
"Suo ",
"Ban ",
"Thoy ",
"Chi ",
"Sang ",
"Niao ",
"Ying ",
"Jie ",
"Ken ",
"Huai ",
"Ku ",
"Lian ",
"Bao ",
"Li ",
"Sup ",
"Shi ",
"Lwu ",
"Yi ",
"Chep ",
"Sel ",
"Xian ",
"Wei ",
"Biao ",
"Cao ",
"Ji ",
"Kang ",
"Sen ",
"Bao ",
"Yang ",
"Chihaya ",
"Pu ",
"Jian ",
"Zhuan ",
"Jian ",
"Zui ",
"Ji ",
"Dan ",
"Cap ",
"Fan ",
"Bo ",
"Xiang ",
"Xin ",
"Bie ",
"Rao ",
"Man ",
"Lan ",
"O ",
"Duo ",
"Koy ",
"Cao ",
"Sui ",
"Nong ",
"Chem ",
"Lian ",
"Pyek ",
"Kum ",
"Tang ",
"Shu ",
"Tan ",
"Bi ",
"Lam ",
"Pu ",
"Yu ",
"Zhi ",
"",
"Shu ",
"Mal ",
"Shi ",
"Bai ",
"Hil ",
"Bo ",
"Chin ",
"Lai ",
"Long ",
"Sup ",
"Xian ",
"Lan ",
"Zhe ",
"Dai ",
"Tasuki ",
"Zan ",
"Shi ",
"Jian ",
"Pan ",
"Yi ",
"Ran ",
"Ya ",
"Se ",
"Xi ",
"Yo ",
"Feng ",
"Tam ",
"",
"Biao ",
"Pok ",
"Phay ",
"Hayk ",
"Ki ",
"Ji ",
"Kyen ",
"Kwan ",
"Bian ",
"Yan ",
"Kyu ",
"Jue ",
"Pian ",
"Mao ",
"Myek ",
"Mi ",
"Mie ",
"Si ",
"Sa ",
"Cem ",
"Luo ",
"Kak ",
"Mi ",
"Tiao ",
"Lian ",
"Yao ",
"Zhi ",
"Jun ",
"Kyek ",
"Shan ",
"Wei ",
"Xi ",
"Tian ",
"Yu ",
"Lam ",
"Ak ",
"To ",
"Chin ",
"Pang ",
"Ki ",
"Ming ",
"Ying ",
"Kwu ",
"Qu ",
"Zhan ",
"Kun ",
"Kwan ",
"Deng ",
"Jian ",
"Luo ",
"Che ",
"Jian ",
"Wei ",
"Kak ",
"Qu ",
"Luo ",
"Lam ",
"Shen ",
"Cek ",
"Kwan ",
"Jian ",
"Guan ",
"Yan ",
"Gui ",
"Mi ",
"Shi ",
"Zhan ",
"Lan ",
"Jue ",
"Ji ",
"Xi ",
"Di ",
"Tian ",
"Yu ",
"Gou ",
"Jin ",
"Qu ",
"Kak ",
"Jiu ",
"Kun ",
"Cu ",
"Kyel ",
"Zhi ",
"Chao ",
"Ji ",
"Ko ",
"Dan ",
"Ca ",
"Ce ",
"Shang ",
"Hua ",
"Quan ",
"Ge ",
"Chi ",
"Hay ",
"Gui ",
"Koyng ",
"Chok ",
"Jie ",
"Hun ",
"Qiu ",
"Xing ",
"Sok ",
"Ni ",
"Ji ",
"Lu ",
"Zhi ",
"Zha ",
"Phil ",
"Xing ",
"Kok ",
"Sang ",
"Gong ",
"Chi ",
"Xue ",
"Chok ",
"Xi ",
"Yi ",
"Lu ",
"Jue ",
"Xi ",
"Yan ",
"Xi ",
],
"x8a": [
"En ",
"Gonben ",
"Ceng ",
"Pwu ",
"Qiu ",
"Qiu ",
"Jiao ",
"Koyng ",
"Kyey ",
"Pem ",
"Sin ",
"Diao ",
"Hong ",
"Cha ",
"Tho ",
"Xu ",
"Al ",
"I ",
"Ren ",
"Hwun ",
"Un ",
"San ",
"Hul ",
"Thak ",
"Ki ",
"Xun ",
"Un ",
"Wa ",
"Fen ",
"A ",
"Yao ",
"Song ",
"Shen ",
"Yin ",
"Hun ",
"Kyel ",
"Xiao ",
"Nwul ",
"Chen ",
"You ",
"Zhi ",
"Xiong ",
"Pang ",
"Xin ",
"Chao ",
"Sel ",
"Xian ",
"Sha ",
"Tun ",
"He ",
"Yi ",
"Yek ",
"So ",
"Chi ",
"Ka ",
"Shen ",
"He ",
"Xu ",
"Cin ",
"Cwu ",
"Ceng ",
"Gou ",
"Ca ",
"Zi ",
"Zhan ",
"Ko ",
"Fu ",
"Quan ",
"Die ",
"Ling ",
"Ce ",
"Yang ",
"Li ",
"Nao ",
"Pan ",
"Cwu ",
"Gan ",
"Yi ",
"Ke ",
"Ao ",
"Sa ",
"I ",
"I ",
"Qu ",
"Co ",
"Phyeng ",
"Phi ",
"Hyeng ",
"Kwul ",
"Ba ",
"Da ",
"Ce ",
"Tao ",
"Zhu ",
"Sa ",
"Zhe ",
"Yeng ",
"Hwu ",
"Swun ",
"Yey ",
"Huang ",
"He ",
"Si ",
"Cha ",
"Jiao ",
"Si ",
"Hen ",
"Tha ",
"Kwu ",
"Kwey ",
"Cen ",
"Hui ",
"Hil ",
"Hwa ",
"Hay ",
"Sang ",
"Wei ",
"Sen ",
"Chou ",
"Tong ",
"Mi ",
"Chem ",
"Ming ",
"E ",
"Hoy ",
"Yan ",
"Xiong ",
"Kway ",
"Er ",
"Beng ",
"Co ",
"Chi ",
"Loy ",
"Cwu ",
"Kwang ",
"Kwa ",
"Wu ",
"Yey ",
"Teng ",
"Ji ",
"Ci ",
"In ",
"Su ",
"Lang ",
"E ",
"Kwang ",
"Huy ",
"Se ",
"Ting ",
"Than ",
"Bo ",
"Chan ",
"Yu ",
"Heng ",
"Cho ",
"Qin ",
"Shua ",
"An ",
"E ",
"Xiao ",
"Seng ",
"Kyey ",
"Hyen ",
"Mwu ",
"O ",
"Ko ",
"Song ",
"Pho ",
"Hoy ",
"Jing ",
"Sel ",
"Zhen ",
"Sel ",
"Tok ",
"Yasashi ",
"Chang ",
"Swu ",
"Jie ",
"Kwa ",
"Qu ",
"Cong ",
"Xiao ",
"Swu ",
"Wang ",
"Xuan ",
"Pi ",
"Chi ",
"Ta ",
"Uy ",
"Na ",
"Un ",
"Co ",
"Pi ",
"Chuo ",
"Chem ",
"Chen ",
"Swun ",
"Ji ",
"Qi ",
"Tam ",
"Zhui ",
"Wi ",
"Ju ",
"Cheng ",
"Jian ",
"Cayng ",
"Ze ",
"Chwu ",
"Qian ",
"Chak ",
"Lyang ",
"Jian ",
"Zhu ",
"Hao ",
"Lon ",
"Sim ",
"Biao ",
"Huai ",
"Pian ",
"Yu ",
"Chep ",
"Se ",
"Phyen ",
"Si ",
"Hwen ",
"Si ",
"Wen ",
"Hua ",
"Ak ",
"Zhong ",
"Chey ",
"Hay ",
"Fu ",
"Pu ",
"Ceng ",
"Kan ",
"Qi ",
"Yu ",
"Ca ",
"Chuan ",
"Si ",
"Hwi ",
"Yin ",
"Am ",
"Ham ",
"Nan ",
"Sim ",
"Phwung ",
"Cey ",
"Yang ",
"En ",
"Heng ",
"Hwen ",
"Ge ",
"Nak ",
"Qi ",
],
"x8b": [
"Mo ",
"Al ",
"Wi ",
"",
"Tung ",
"Zou ",
"Shan ",
"Ken ",
"Bo ",
"",
"Huang ",
"Huo ",
"Ka ",
"Ying ",
"Mi ",
"Xiao ",
"Mil ",
"Hyey ",
"Qiang ",
"Chen ",
"Hak ",
"Ti ",
"So ",
"Pang ",
"Chi ",
"Kyem ",
"Ik ",
"Kang ",
"Wen ",
"Sa ",
"Xue ",
"Tao ",
"Yo ",
"Yao ",
"",
"Wu ",
"Biao ",
"Cong ",
"Kyeng ",
"Li ",
"Mo ",
"Mo ",
"Shang ",
"Cek ",
"Lyu ",
"Cen ",
"Ze ",
"Jie ",
"Lian ",
"Lou ",
"Can ",
"Kwu ",
"Guan ",
"Xi ",
"Zhuo ",
"O ",
"Ao ",
"Kun ",
"Zhe ",
"Yi ",
"Hu ",
"Jiang ",
"Man ",
"Chao ",
"Han ",
"Hwa ",
"Chan ",
"Hwu ",
"Zeng ",
"Se ",
"Xi ",
"She ",
"Dui ",
"Cung ",
"Nao ",
"Lan ",
"Wa ",
"Ying ",
"Hyul ",
"Ki ",
"Zun ",
"Jiao ",
"Bo ",
"Hui ",
"Sen ",
"Mu ",
"Cham ",
"Zha ",
"Sik ",
"Cho ",
"Tam ",
"Zen ",
"Po ",
"Sheng ",
"Xuan ",
"Co ",
"Tan ",
"Dang ",
"Sui ",
"Qian ",
"Ji ",
"Jiao ",
"Kyeng ",
"Lian ",
"Nou ",
"Yi ",
"Ai ",
"Sem ",
"Pi ",
"Hui ",
"Hua ",
"Yek ",
"Uy ",
"Sen ",
"Rang ",
"Nou ",
"Kyen ",
"Zhui ",
"Ta ",
"Ho ",
"Zhou ",
"Hao ",
"Ye ",
"Ying ",
"Jian ",
"Yey ",
"Jian ",
"Hyey ",
"Tok ",
"Zhe ",
"Xuan ",
"Chan ",
"Lei ",
"Shen ",
"Wei ",
"Chan ",
"Li ",
"Yu ",
"Pyen ",
"Zhe ",
"Yen ",
"E ",
"Swu ",
"Wei ",
"Swu ",
"Yao ",
"Cham ",
"Yang ",
"Yin ",
"Lan ",
"Cham ",
"Huo ",
"Zhe ",
"Hwan ",
"Chan ",
"Yi ",
"Tang ",
"Zhan ",
"En ",
"Du ",
"Yan ",
"Ji ",
"Ding ",
"Fu ",
"Ren ",
"Ji ",
"Jie ",
"Hong ",
"Tao ",
"Rang ",
"Shan ",
"Qi ",
"Thak ",
"Xun ",
"Yi ",
"Xun ",
"Ji ",
"Ren ",
"Jiang ",
"Hui ",
"Ou ",
"Ju ",
"Ya ",
"Ne ",
"Xu ",
"E ",
"Lun ",
"Xiong ",
"Song ",
"Feng ",
"She ",
"Fang ",
"Jue ",
"Zheng ",
"Gu ",
"He ",
"Ping ",
"Zu ",
"Shi ",
"Xiong ",
"Zha ",
"Su ",
"Zhen ",
"Di ",
"Zou ",
"Ci ",
"Qu ",
"Zhao ",
"Bi ",
"Yi ",
"Yi ",
"Kuang ",
"Lei ",
"Shi ",
"Gua ",
"Shi ",
"Jie ",
"Hui ",
"Cheng ",
"Zhu ",
"Shen ",
"Hua ",
"Dan ",
"Gou ",
"Quan ",
"Gui ",
"Xun ",
"Yi ",
"Zheng ",
"Gai ",
"Xiang ",
"Cha ",
"Hun ",
"Xu ",
"Zhou ",
"Jie ",
"Wu ",
"Yu ",
"Qiao ",
"Wu ",
"Gao ",
"You ",
"Hui ",
"Kuang ",
"Shuo ",
"Song ",
"Ai ",
"Qing ",
"Zhu ",
"Zou ",
"Nuo ",
"Du ",
"Zhuo ",
"Fei ",
"Ke ",
"Wei ",
],
"x8c": [
"Yu ",
"Shui ",
"Shen ",
"Diao ",
"Chan ",
"Liang ",
"Zhun ",
"Sui ",
"Tan ",
"Shen ",
"Yi ",
"Mou ",
"Chen ",
"Die ",
"Huang ",
"Jian ",
"Xie ",
"Nue ",
"Ye ",
"Wei ",
"E ",
"Yu ",
"Xuan ",
"Chan ",
"Zi ",
"An ",
"Yan ",
"Di ",
"Mi ",
"Pian ",
"Se ",
"Mo ",
"Dang ",
"Su ",
"Xie ",
"Yao ",
"Bang ",
"Shi ",
"Qian ",
"Mi ",
"Jin ",
"Man ",
"Zhe ",
"Jian ",
"Miu ",
"Tan ",
"Zen ",
"Qiao ",
"Lan ",
"Pu ",
"Jue ",
"Yan ",
"Qian ",
"Zhan ",
"Chen ",
"Kok ",
"Qian ",
"Hong ",
"Xia ",
"Jue ",
"Hong ",
"Han ",
"Hong ",
"Kyey ",
"Xi ",
"Hwal ",
"Liao ",
"Han ",
"Du ",
"Long ",
"Twu ",
"Kang ",
"Kay ",
"Si ",
"Lyey ",
"Deng ",
"Wan ",
"Bi ",
"Swu ",
"Xian ",
"Phwung ",
"Zhi ",
"Zhi ",
"Yan ",
"Yan ",
"Si ",
"Chwuk ",
"Hui ",
"Tun ",
"Yi ",
"Ton ",
"Yi ",
"Jian ",
"Ba ",
"Hou ",
"E ",
"Cu ",
"Sang ",
"Hwan ",
"Yen ",
"Ken ",
"Gai ",
"Qu ",
"Fu ",
"Huy ",
"Bin ",
"Ho ",
"Yey ",
"Ce ",
"Jia ",
"Fen ",
"Xi ",
"Bo ",
"Wen ",
"Huan ",
"Pin ",
"Di ",
"Zong ",
"Fen ",
"Yi ",
"Chi ",
"Phyo ",
"Si ",
"An ",
"Pi ",
"Na ",
"Pi ",
"Gou ",
"Na ",
"You ",
"Cho ",
"Mayk ",
"Si ",
"Hyu ",
"Huan ",
"Kun ",
"He ",
"He ",
"Mayk ",
"Han ",
"Mo ",
"Li ",
"Ni ",
"Bi ",
"Yu ",
"Jia ",
"Tuan ",
"Myo ",
"Pi ",
"Xi ",
"E ",
"Ju ",
"Mayk ",
"Chu ",
"Tan ",
"Huan ",
"Jue ",
"Phay ",
"Ceng ",
"Yuan ",
"Pwu ",
"Cay ",
"Kong ",
"Te ",
"Yi ",
"Hang ",
"Wan ",
"Pin ",
"Hwa ",
"Phan ",
"Tham ",
"Kwan ",
"Chayk ",
"Zhi ",
"Er ",
"Ce ",
"Sey ",
"Bi ",
"Ca ",
"I ",
"Kwi ",
"Pian ",
"Phyem ",
"May ",
"Tay ",
"Sheng ",
"Hwang ",
"Pi ",
"Chep ",
"I ",
"Chi ",
"Mwu ",
"Ha ",
"Pwun ",
"Loy ",
"Im ",
"Hoy ",
"Gai ",
"Pyen ",
"Ca ",
"Ka ",
"Xu ",
"Cek ",
"Jiao ",
"Gai ",
"Zang ",
"Jian ",
"Ying ",
"Xun ",
"Cin ",
"She ",
"Pin ",
"Pin ",
"Qiu ",
"She ",
"Chuan ",
"Zang ",
"Cwu ",
"Loy ",
"Chan ",
"Sa ",
"Chen ",
"Sang ",
"Tian ",
"Pay ",
"Kayng ",
"Hyen ",
"May ",
"Chen ",
"Sui ",
"Pwu ",
"Tan ",
"Cong ",
"Cong ",
"Cil ",
"Ji ",
"Cang ",
"To ",
"Jin ",
"Xiong ",
"Shun ",
"Yun ",
"Bao ",
"Zai ",
"Loy ",
"Feng ",
"Cang ",
"Ji ",
"Sheng ",
"Ai ",
"Zhuan ",
"Pwu ",
"Kwu ",
"Say ",
"Sayk ",
"Liao ",
],
"x8d": [
"Wei ",
"Bai ",
"Chen ",
"Zhuan ",
"Ci ",
"Chwey ",
"Biao ",
"Yun ",
"Cung ",
"Tan ",
"Chan ",
"An ",
"",
"Sem ",
"Wan ",
"Yeng ",
"Sin ",
"Gan ",
"Xian ",
"Cang ",
"Pi ",
"Du ",
"Sok ",
"Yan ",
"",
"Xuan ",
"Long ",
"Kong ",
"Cang ",
"Bei ",
"Zhen ",
"Fu ",
"Yuan ",
"Gong ",
"Cai ",
"Ze ",
"Xian ",
"Bai ",
"Zhang ",
"Huo ",
"Zhi ",
"Fan ",
"Tan ",
"Pin ",
"Bian ",
"Gou ",
"Zhu ",
"Guan ",
"Er ",
"Jian ",
"Bi ",
"Shi ",
"Tie ",
"Gui ",
"Kuang ",
"Dai ",
"Mao ",
"Fei ",
"He ",
"Yi ",
"Zei ",
"Zhi ",
"Jia ",
"Hui ",
"Zi ",
"Ren ",
"Lu ",
"Zang ",
"Zi ",
"Gai ",
"Jin ",
"Qiu ",
"Zhen ",
"Lai ",
"She ",
"Fu ",
"Du ",
"Ji ",
"Shu ",
"Shang ",
"Si ",
"Bi ",
"Cwu ",
"Geng ",
"Pei ",
"Tan ",
"Lai ",
"Feng ",
"Zhui ",
"Fu ",
"Zhuan ",
"Sai ",
"Ze ",
"Yan ",
"Zan ",
"Yun ",
"Zeng ",
"Shan ",
"Ying ",
"Gan ",
"Cek ",
"Xi ",
"Sa ",
"Nan ",
"Xiong ",
"Xi ",
"Cheng ",
"Hyek ",
"Cheng ",
"Ca ",
"Xia ",
"Tang ",
"Cwu ",
"Zou ",
"Li ",
"Kyu ",
"Pwu ",
"Zhao ",
"Kan ",
"Ki ",
"Shan ",
"Qiong ",
"Qin ",
"Xian ",
"Ci ",
"Jue ",
"Qin ",
"Chi ",
"Ci ",
"Cin ",
"Chen ",
"Die ",
"Ju ",
"Cho ",
"Di ",
"Se ",
"Zhan ",
"Zhu ",
"Wel ",
"Qu ",
"Jie ",
"Chi ",
"Cwu ",
"Gua ",
"Hyel ",
"Ca ",
"Tiao ",
"Duo ",
"Lie ",
"Gan ",
"Suo ",
"Cu ",
"Xi ",
"Co ",
"Su ",
"Yin ",
"Ju ",
"Jian ",
"Que ",
"Tang ",
"Cho ",
"Cui ",
"Lu ",
"Chwi ",
"Dang ",
"Qiu ",
"Zi ",
"Ti ",
"Chwu ",
"Chi ",
"Huang ",
"Kyo ",
"Qiao ",
"Yao ",
"Zao ",
"Ti ",
"",
"Zan ",
"Chan ",
"Cok ",
"Pa ",
"Pak ",
"Ku ",
"Ke ",
"Dun ",
"Jue ",
"Pwu ",
"Chen ",
"Kyen ",
"Fang ",
"Ci ",
"Sap ",
"Yue ",
"Pa ",
"Ki ",
"Yue ",
"Qiang ",
"Tuo ",
"Thay ",
"Yi ",
"Nian ",
"Ling ",
"Mei ",
"Pal ",
"Cil ",
"Ku ",
"Tha ",
"Ka ",
"Ci ",
"Pao ",
"Qia ",
"Zhu ",
"Ju ",
"Cep ",
"Chek ",
"Pwu ",
"Pan ",
"Ju ",
"San ",
"Pha ",
"Ni ",
"Ke ",
"Li ",
"Kun ",
"Yi ",
"Cek ",
"Dai ",
"Sen ",
"Kyo ",
"Duo ",
"Zhu ",
"Quan ",
"Kwa ",
"Zhuai ",
"Kwey ",
"Kong ",
"Kyu ",
"Xiang ",
"Chi ",
"Lo ",
"Beng ",
"Chi ",
"Jia ",
"Co ",
"Cai ",
"Chen ",
"Ta ",
"Qiao ",
"Bi ",
"Xian ",
"Duo ",
"Ji ",
"Kwuk ",
"Ki ",
"Shu ",
"Tu ",
],
"x8e": [
"Chu ",
"Jing ",
"Nie ",
"Xiao ",
"Bo ",
"Chi ",
"Cwun ",
"Mou ",
"So ",
"Lyang ",
"Yong ",
"Jiao ",
"Chou ",
"Qiao ",
"Mau ",
"Tap ",
"Chen ",
"Qi ",
"Wo ",
"Wei ",
"Thak ",
"Jie ",
"Ji ",
"Nie ",
"Ju ",
"Ju ",
"Lun ",
"Lu ",
"Leng ",
"Kwa ",
"Ke ",
"Ci ",
"Wan ",
"Quan ",
"Ti ",
"Pwu ",
"Zu ",
"Qie ",
"Ji ",
"Cu ",
"Zong ",
"Chay ",
"Cong ",
"Peng ",
"Zhi ",
"Zheng ",
"Dian ",
"Zhi ",
"Yu ",
"Thak ",
"Dun ",
"Chun ",
"Yong ",
"Cong ",
"Cey ",
"Zhe ",
"Chen ",
"Tan ",
"Jian ",
"Gua ",
"Tang ",
"Wu ",
"Fu ",
"Zu ",
"Cep ",
"Phyen ",
"Yu ",
"Nuo ",
"Cey ",
"Cha ",
"Tui ",
"Ken ",
"To ",
"Cha ",
"Hyey ",
"Ta ",
"Chang ",
"Zhan ",
"Dian ",
"Cey ",
"Ji ",
"Nie ",
"Man ",
"Liu ",
"Zhan ",
"Phil ",
"Chong ",
"Lu ",
"Liao ",
"Chwuk ",
"Tang ",
"Dai ",
"Chwuk ",
"Xi ",
"Kui ",
"Cek ",
"Chek ",
"Qiang ",
"Di ",
"Man ",
"Cong ",
"Lian ",
"Beng ",
"Zao ",
"Nian ",
"Pyel ",
"Tui ",
"Ju ",
"Deng ",
"Chung ",
"Xian ",
"Pen ",
"Cwu ",
"Zhong ",
"Cwun ",
"Bo ",
"Chwuk ",
"Zu ",
"Kwel ",
"Jue ",
"Lin ",
"Ta ",
"Kyo ",
"Qiao ",
"Pok ",
"Liao ",
"Dun ",
"Cuan ",
"Kuang ",
"Co ",
"Ta ",
"Bi ",
"Bi ",
"Chok ",
"Ju ",
"Ce ",
"Qiao ",
"Ton ",
"Cwu ",
"Cey ",
"Wu ",
"Yak ",
"Nian ",
"Lin ",
"Lyep ",
"Chek ",
"Lyek ",
"Ci ",
"Cen ",
"Chu ",
"Duan ",
"Wei ",
"Long ",
"Lin ",
"Sen ",
"Wei ",
"Zuan ",
"Lan ",
"Sep ",
"Rang ",
"Xie ",
"Sep ",
"Ta ",
"Qu ",
"Jie ",
"Cuan ",
"Zuan ",
"Xi ",
"Kui ",
"Kwak ",
"Lin ",
"Sin ",
"Kwung ",
"Dan ",
"Segare ",
"Qu ",
"Ti ",
"Tha ",
"Duo ",
"Kwung ",
"Lang ",
"Nerau ",
"Luo ",
"Ai ",
"Ji ",
"Ju ",
"Tang ",
"Utsuke ",
"",
"Yan ",
"Shitsuke ",
"Kang ",
"Kwu ",
"Lou ",
"Lao ",
"Tuo ",
"Zhi ",
"Yagate ",
"Chey ",
"Dao ",
"Yagate ",
"Yu ",
"Cha ",
"Al ",
"Kwey ",
"Kwun ",
"Wei ",
"Yue ",
"Xin ",
"Di ",
"Hen ",
"Fan ",
"Ren ",
"Shan ",
"Qiang ",
"Shu ",
"Tun ",
"Chen ",
"Dai ",
"E ",
"Na ",
"Qi ",
"Mao ",
"Yen ",
"Ren ",
"Fan ",
"Cen ",
"Hong ",
"Hu ",
"Qu ",
"Huang ",
"Di ",
"Lyeng ",
"Dai ",
"Ao ",
"Cin ",
"Fan ",
"Kuang ",
"Ang ",
"Peng ",
"Bei ",
"Gu ",
"Gu ",
"Pao ",
"Zhu ",
"Rong ",
"E ",
"Ba ",
"Chwuk ",
"Ci ",
"Cho ",
"Ka ",
"Il ",
"Kyeng ",
"Sik ",
"Pyeng ",
],
"x8f": [
"Er ",
"Qiong ",
"Ju ",
"Kyo ",
"Guang ",
"Lo ",
"Kai ",
"Cen ",
"Cwu ",
"Cay ",
"Zhi ",
"She ",
"Liang ",
"Yu ",
"Shao ",
"You ",
"Huan ",
"Yun ",
"Chep ",
"Man ",
"Po ",
"Kyeng ",
"Zhou ",
"Ni ",
"Ling ",
"Zhe ",
"Zhan ",
"Lyang ",
"Chi ",
"Hwi ",
"Mang ",
"Chel ",
"Guo ",
"Kan ",
"Yi ",
"Peng ",
"Qian ",
"Gun ",
"Lyen ",
"Pyeng ",
"Kwan ",
"Pay ",
"Lyun ",
"Pai ",
"Liang ",
"Ruan ",
"Rou ",
"Cip ",
"Yang ",
"Xian ",
"Chuan ",
"Cwu ",
"Chun ",
"Ge ",
"You ",
"Hong ",
"Swu ",
"Pok ",
"Zi ",
"Pok ",
"Wen ",
"Ben ",
"Cen ",
"Ye ",
"On ",
"Tao ",
"Kok ",
"Zhen ",
"Hal ",
"Wen ",
"Lok ",
"Jiu ",
"Chao ",
"Cen ",
"Wei ",
"Hun ",
"Sori ",
"Chel ",
"Kyo ",
"Zhan ",
"Pu ",
"Lao ",
"Pwun ",
"Fan ",
"Lin ",
"Ge ",
"Se ",
"Kem ",
"Hwan ",
"Yi ",
"Ji ",
"Dui ",
"Er ",
"Ye ",
"Ham ",
"Koyng ",
"Lei ",
"Pi ",
"Lyek ",
"Lyek ",
"Lo ",
"Lin ",
"Che ",
"Ya ",
"Gui ",
"Xuan ",
"Di ",
"Ren ",
"Zhuan ",
"E ",
"Lun ",
"Ruan ",
"Hong ",
"Gu ",
"Ke ",
"Lu ",
"Zhou ",
"Zhi ",
"Yi ",
"Hu ",
"Zhen ",
"Li ",
"Yao ",
"Qing ",
"Shi ",
"Zai ",
"Zhi ",
"Jiao ",
"Zhou ",
"Quan ",
"Lu ",
"Jiao ",
"Zhe ",
"Fu ",
"Liang ",
"Nian ",
"Bei ",
"Hui ",
"Gun ",
"Wang ",
"Liang ",
"Chuo ",
"Zi ",
"Cou ",
"Fu ",
"Ji ",
"Wen ",
"Shu ",
"Pei ",
"Yuan ",
"Xia ",
"Zhan ",
"Lu ",
"Che ",
"Lin ",
"Sin ",
"Ko ",
"Ci ",
"Sa ",
"Phi ",
"Zui ",
"Bian ",
"Lal ",
"Lal ",
"Ci ",
"Xue ",
"Phan ",
"Bian ",
"Pyen ",
"Bian ",
"",
"Bian ",
"Ban ",
"Sa ",
"Pyen ",
"Pyen ",
"Cin ",
"Yok ",
"Nong ",
"Nong ",
"Zhen ",
"Chuo ",
"Chuo ",
"Suberu ",
"Reng ",
"Pyen ",
"Pyen ",
"Sip ",
"Ip ",
"Liao ",
"Da ",
"Chen ",
"Kan ",
"Qian ",
"O ",
"O ",
"Hul ",
"Sin ",
"Yi ",
"Guo ",
"Mai ",
"Qi ",
"Za ",
"Wang ",
"tu ",
"Twun ",
"Yeng ",
"Ti ",
"Yun ",
"Kun ",
"Hang ",
"A ",
"Pan ",
"O ",
"Da ",
"E ",
"Huan ",
"Zhe ",
"Totemo ",
"Jin ",
"Yuan ",
"Wei ",
"Lian ",
"Chi ",
"Che ",
"Ni ",
"Cho ",
"Zhi ",
"I ",
"Hyeng ",
"Ka ",
"Chen ",
"Thay ",
"I ",
"Cek ",
"Pak ",
"Wang ",
"Cil ",
"Ze ",
"Tao ",
"Swul ",
"Tuo ",
"Kep ",
"Jing ",
"Hoy ",
"Tong ",
"You ",
"Mi ",
"Pyeng ",
"Cek ",
"Nay ",
"Yi ",
"Jie ",
"Chwu ",
"Lie ",
"Xun ",
],
"x90": [
"Thoy ",
"Song ",
"Kwal ",
"To ",
"Pang ",
"Hwu ",
"Yek ",
"Ci ",
"Hyeng ",
"Xuan ",
"Xun ",
"Pho ",
"Yu ",
"So ",
"Qiu ",
"Thwu ",
"Chwuk ",
"Kwu ",
"Di ",
"Di ",
"To ",
"Kyeng ",
"Cek ",
"Twu ",
"Yi ",
"Ce ",
"Thong ",
"Guang ",
"Wu ",
"Se ",
"Lyeng ",
"Sok ",
"Co ",
"Cwun ",
"Pong ",
"Lyen ",
"Suo ",
"Hoy ",
"Li ",
"Sako ",
"Lai ",
"Ben ",
"Cuo ",
"Jue ",
"Beng ",
"Huan ",
"Chey ",
"Lu ",
"Yu ",
"Cwu ",
"Cin ",
"Yu ",
"Thak ",
"Kyu ",
"Wi ",
"Ti ",
"Il ",
"Da ",
"Wen ",
"Luo ",
"Phip ",
"Nuo ",
"Yu ",
"Dang ",
"Sui ",
"Twun ",
"Swu ",
"Yan ",
"Chuan ",
"Ci ",
"Ti ",
"Wu ",
"Shi ",
"Ceng ",
"Yu ",
"Wun ",
"E ",
"Phyen ",
"Kwa ",
"Al ",
"Ha ",
"Hwang ",
"Cwu ",
"To ",
"Tal ",
"Wi ",
"Appare ",
"Yi ",
"Kwu ",
"Yo ",
"Chu ",
"Lyu ",
"Son ",
"Tap ",
"Chey ",
"Ci ",
"Wen ",
"So ",
"Ta ",
"Kyen ",
"",
"Yao ",
"Guan ",
"Zhang ",
"O ",
"Cek ",
"Ce ",
"Chi ",
"Sok ",
"Co ",
"Cha ",
"Twun ",
"Di ",
"Lou ",
"Ci ",
"Cuo ",
"Lin ",
"Cwun ",
"Yo ",
"Chen ",
"Sen ",
"Yu ",
"Yu ",
"Wu ",
"Lyo ",
"Ke ",
"Shi ",
"Phi ",
"Yo ",
"May ",
"Hay ",
"Swu ",
"Hwan ",
"Cen ",
"Teng ",
"I ",
"Mak ",
"Bian ",
"Pyen ",
"La ",
"Lye ",
"Yuan ",
"Yao ",
"La ",
"Li ",
"Up ",
"Ting ",
"Deng ",
"Qi ",
"Ong ",
"Shan ",
"Han ",
"Yu ",
"Mang ",
"Ru ",
"Kong ",
"",
"Kuang ",
"Fu ",
"Kang ",
"Pin ",
"Pang ",
"Hyeng ",
"Na ",
"",
"Shen ",
"Pang ",
"Yuan ",
"Chon ",
"Huo ",
"Sa ",
"Bang ",
"Wu ",
"Ju ",
"You ",
"Han ",
"Thay ",
"Kwu ",
"Bi ",
"Pi ",
"Pyeng ",
"So ",
"Phay ",
"Wa ",
"Ce ",
"Zou ",
"Ye ",
"Lin ",
"Kuang ",
"Kyu ",
"Cwu ",
"Shi ",
"Ku ",
"Wuk ",
"Gai ",
"Hap ",
"Kuk ",
"Cil ",
"Ji ",
"Xun ",
"Hou ",
"Xing ",
"Kyo ",
"Xi ",
"Gui ",
"Nuo ",
"Lang ",
"Jia ",
"Kuai ",
"Zheng ",
"Rou ",
"Yun ",
"Yan ",
"Cheng ",
"Dou ",
"Chi ",
"Lu ",
"Fu ",
"Wu ",
"Fu ",
"Ko ",
"Hak ",
"Lang ",
"Kyep ",
"Geng ",
"Kwun ",
"Yeng ",
"Bo ",
"Kuk ",
"Bei ",
"Li ",
"Yun ",
"Pwu ",
"Xiao ",
"Che ",
"Pi ",
"Qing ",
"Kwak ",
"",
"Tam ",
"Zou ",
"Ping ",
"Lai ",
"Ni ",
"Chim ",
"Wu ",
"Bu ",
"Hyang ",
"Dan ",
"Ju ",
"Yong ",
"Qiao ",
"Yi ",
"To ",
"Yan ",
"Mei ",
],
"x91": [
"Ruo ",
"Bei ",
"Ak ",
"Yu ",
"Juan ",
"Yu ",
"Yun ",
"Hou ",
"Kui ",
"Hyang ",
"Xiang ",
"Sou ",
"Tang ",
"Ming ",
"Xi ",
"Ru ",
"Chu ",
"Zi ",
"Chwu ",
"Ju ",
"O ",
"Hyang ",
"Yun ",
"Hao ",
"Yong ",
"Pi ",
"Mo ",
"Chao ",
"Fu ",
"Liao ",
"Un ",
"Zhuan ",
"Hu ",
"Qiao ",
"En ",
"Cang ",
"Man ",
"Qiao ",
"Xu ",
"Tung ",
"Bi ",
"Xin ",
"Bi ",
"Ceng ",
"Wei ",
"Ceng ",
"Mao ",
"Shan ",
"Lin ",
"Pha ",
"Tan ",
"Meng ",
"Ep ",
"Cao ",
"Hoy ",
"Feng ",
"Meng ",
"Zou ",
"Kwang ",
"Lian ",
"Zan ",
"Cen ",
"You ",
"Qi ",
"Yan ",
"Chan ",
"Zan ",
"Ling ",
"Huan ",
"Xi ",
"Feng ",
"Zan ",
"Lyek ",
"Yu ",
"Ceng ",
"Chwu ",
"Cak ",
"Pay ",
"Cwu ",
"Yi ",
"Hang ",
"Yu ",
"Cwu ",
"Yan ",
"Zui ",
"Mao ",
"Tham ",
"Hwu ",
"Tou ",
"Zhen ",
"Fen ",
"Sakenomoto ",
"",
"Yun ",
"Tai ",
"Tian ",
"Qia ",
"Tuo ",
"Cho ",
"Kem ",
"Ko ",
"So ",
"Pal ",
"Chou ",
"Zai ",
"Myeng ",
"Lak ",
"Chuo ",
"Swu ",
"You ",
"Tong ",
"Zhi ",
"Xian ",
"Cang ",
"Ceng ",
"Yin ",
"To ",
"Hyo ",
"May ",
"Hok ",
"San ",
"Loy ",
"Pu ",
"Zui ",
"Hai ",
"Yan ",
"Xi ",
"Niang ",
"Wei ",
"Lu ",
"Lan ",
"Em ",
"Tao ",
"Pei ",
"Zhan ",
"Swun ",
"Tan ",
"Chwi ",
"Chuo ",
"Cho ",
"Kun ",
"Cey ",
"Mian ",
"Du ",
"Ho ",
"Xu ",
"Seng ",
"Tan ",
"Jiu ",
"Chun ",
"On ",
"Pal ",
"Ke ",
"Sou ",
"Mi ",
"Quan ",
"Chwu ",
"Cuo ",
"On ",
"Yong ",
"Ang ",
"Zha ",
"Hay ",
"Tang ",
"Cang ",
"Piao ",
"Shan ",
"Yu ",
"Li ",
"Zao ",
"Lyo ",
"Uy ",
"Cang ",
"Bu ",
"Cho ",
"Hyey ",
"Tan ",
"Pal ",
"Nong ",
"Yi ",
"Lyey ",
"Kyak ",
"Jiao ",
"Yi ",
"Niang ",
"Ru ",
"Xun ",
"Chou ",
"Yan ",
"Ling ",
"Mi ",
"Mi ",
"Yang ",
"Hun ",
"Jiao ",
"Si ",
"Mi ",
"Yem ",
"Pyen ",
"Chay ",
"Sek ",
"Yu ",
"Shi ",
"Sek ",
"Li ",
"Cwung ",
"Ya ",
"Lyang ",
"Li ",
"Kim ",
"",
"Ga ",
"Yi ",
"Liao ",
"Dao ",
"Soy ",
"Ceng ",
"Po ",
"Qiu ",
"He ",
"Pwu ",
"Chim ",
"Zhi ",
"Ba ",
"Luan ",
"Fu ",
"Nai ",
"Co ",
"Sam ",
"Qiao ",
"Kwu ",
"Chen ",
"Zi ",
"Fan ",
"Wu ",
"Hua ",
"Han ",
"Kong ",
"Qi ",
"Mang ",
"Ri ",
"Di ",
"Si ",
"Xi ",
"Yi ",
"Chay ",
"Shi ",
"Tu ",
"Xi ",
"Nu ",
"Qian ",
"Kyuu ",
"Jian ",
"Pi ",
"Ye ",
"Kun ",
],
"x92": [
"Ba ",
"Fang ",
"Chen ",
"Xing ",
"Dou ",
"Yue ",
"Yan ",
"Pwu ",
"Pi ",
"Na ",
"Xin ",
"E ",
"Jue ",
"Twun ",
"Gou ",
"Yin ",
"Kem ",
"Phan ",
"Sap ",
"Ren ",
"Cho ",
"Niu ",
"Fen ",
"Yun ",
"Ji ",
"Qin ",
"Pi ",
"Guo ",
"Hoyng ",
"Yin ",
"Kyun ",
"Co ",
"Yi ",
"Zhong ",
"Nie ",
"Gai ",
"Ri ",
"Huo ",
"Tai ",
"Kang ",
"Habaki ",
"Ro ",
"Ngaak ",
"",
"Duo ",
"Zi ",
"Ni ",
"Tu ",
"Shi ",
"Min ",
"Gu ",
"Ke ",
"Lyeng ",
"Pyeng ",
"Yi ",
"Ko ",
"Pal ",
"Pi ",
"Ok ",
"Si ",
"Zuo ",
"Bu ",
"You ",
"Cen ",
"Kap ",
"Cin ",
"Shi ",
"Shi ",
"Chel ",
"Ke ",
"Chan ",
"Shi ",
"Shi ",
"Hyen ",
"Zhao ",
"Pho ",
"He ",
"Bi ",
"Sayng ",
"Se ",
"Sek ",
"Pak ",
"Cwu ",
"Chi ",
"Za ",
"Po ",
"Tong ",
"Kyem ",
"Fu ",
"Zhai ",
"Liu ",
"Yen ",
"Fu ",
"Li ",
"Wel ",
"Pi ",
"Yang ",
"Ban ",
"Pal ",
"Jie ",
"Kwu ",
"Swul ",
"Ceng ",
"Mu ",
"Ni ",
"Nie ",
"Di ",
"Jia ",
"Mu ",
"Dan ",
"Shen ",
"Yi ",
"Si ",
"Kwang ",
"Ka ",
"Bei ",
"Jian ",
"Tong ",
"Xing ",
"Hong ",
"Kyo ",
"Chi ",
"Er ",
"Ge ",
"Pyeng ",
"Shi ",
"Mo ",
"Ha ",
"Un ",
"Jun ",
"Zhou ",
"Chong ",
"Shang ",
"Tong ",
"Mo ",
"Lei ",
"Ji ",
"Yu ",
"Xu ",
"Im ",
"Zun ",
"Zhi ",
"Kong ",
"Shan ",
"Chi ",
"Sen ",
"Xing ",
"Cen ",
"Pi ",
"Chel ",
"Swu ",
"Hou ",
"Myeng ",
"Kwa ",
"Co ",
"Sem ",
"Ham ",
"Xiu ",
"Jun ",
"Cha ",
"Lao ",
"Ji ",
"Pi ",
"Ru ",
"Mi ",
"Yi ",
"Yin ",
"Guang ",
"An ",
"Diu ",
"You ",
"Se ",
"Kao ",
"Cen ",
"Luan ",
"Kasugai ",
"Ai ",
"Diao ",
"Han ",
"Yey ",
"Shi ",
"Keng ",
"Kwu ",
"So ",
"Zhe ",
"Swu ",
"Zang ",
"Ti ",
"Cuo ",
"Gua ",
"Gong ",
"Zhong ",
"Dou ",
"Lu ",
"Mei ",
"Lang ",
"Wan ",
"Xin ",
"Yun ",
"Bei ",
"Ok ",
"Su ",
"Yu ",
"Chan ",
"Ceng ",
"Bo ",
"Han ",
"Hyep ",
"Hong ",
"Cen ",
"Pong ",
"Chan ",
"Wan ",
"Zhi ",
"Si ",
"Hyen ",
"Wu ",
"Wu ",
"Tiao ",
"Gong ",
"Zhuo ",
"Lue ",
"Xing ",
"Chim ",
"Shen ",
"Han ",
"Lue ",
"Xie ",
"Se ",
"Ceng ",
"Ju ",
"Xian ",
"Tie ",
"Mang ",
"Pho ",
"Li ",
"Pan ",
"Yey ",
"Cheng ",
"Gao ",
"Li ",
"Te ",
"Pyeng ",
"Zhu ",
"",
"Tu ",
"Liu ",
"Zui ",
"Ke ",
"Cheng ",
"Wen ",
"Jian ",
"Kang ",
"Co ",
"Tao ",
"Chang ",
],
"x93": [
"Lun ",
"Guo ",
"Ling ",
"Bei ",
"Lok ",
"Li ",
"Cheng ",
"Pou ",
"Juan ",
"Min ",
"Zui ",
"Peng ",
"An ",
"Pi ",
"Xian ",
"A ",
"Chwu ",
"Lei ",
"A ",
"Kong ",
"Ta ",
"Kon ",
"Du ",
"Wei ",
"Chwu ",
"Chi ",
"Cayng ",
"Ben ",
"Nie ",
"Cong ",
"Swun ",
"Tam ",
"Ceng ",
"Ki ",
"Cen ",
"Chel ",
"Ki ",
"Yu ",
"Kum ",
"Kwan ",
"Myo ",
"Chang ",
"Cen ",
"Sek ",
"Tong ",
"Tao ",
"Ko ",
"Chak ",
"Shu ",
"Zhen ",
"Lok ",
"Meng ",
"Lu ",
"Hua ",
"Biao ",
"Ga ",
"Lai ",
"Ken ",
"Kazari ",
"Bu ",
"Nai ",
"Wan ",
"Zan ",
"",
"De ",
"Xian ",
"",
"Huo ",
"Liang ",
"",
"Men ",
"Kai ",
"Yeng ",
"Si ",
"Lyen ",
"Kwa ",
"Xian ",
"To ",
"Tu ",
"Wei ",
"Cong ",
"Fu ",
"Rou ",
"Ji ",
"Ak ",
"Rou ",
"Chen ",
"Cey ",
"Zha ",
"Hong ",
"Yang ",
"Tan ",
"Ha ",
"Wu ",
"Keng ",
"Xing ",
"Koyng ",
"Wei ",
"Fu ",
"Zhao ",
"Sap ",
"Qie ",
"She ",
"Hong ",
"Kui ",
"Tian ",
"Mwu ",
"Cho ",
"Cho ",
"Hou ",
"Yu ",
"Cong ",
"Hwan ",
"Ye ",
"Min ",
"Jian ",
"Duan ",
"Ken ",
"Si ",
"Kui ",
"Hu ",
"Xuan ",
"Zhe ",
"Jie ",
"Chim ",
"Bian ",
"Cong ",
"Zi ",
"Xiu ",
"Ye ",
"Mei ",
"Pai ",
"Ai ",
"Jie ",
"",
"Mei ",
"Chuo ",
"Ta ",
"Pang ",
"Xia ",
"Kyem ",
"Suo ",
"Xi ",
"Lyu ",
"Zu ",
"Ye ",
"Nou ",
"Weng ",
"Yong ",
"Tang ",
"Sway ",
"Cayng ",
"Ge ",
"Shuo ",
"Chwu ",
"Pak ",
"Pan ",
"Sa ",
"Bi ",
"Sang ",
"Gang ",
"Ca ",
"Wu ",
"Hyeng ",
"Hwang ",
"Tiao ",
"Liu ",
"Kay ",
"Sun ",
"Sha ",
"Sou ",
"Wan ",
"Ho ",
"Cin ",
"Cin ",
"Luo ",
"Il ",
"Yuan ",
"Tang ",
"Nie ",
"Xi ",
"Jia ",
"Ge ",
"Ma ",
"Juan ",
"Kasugai ",
"Habaki ",
"Suo ",
"",
"",
"",
"Na ",
"Lu ",
"Sway ",
"Ou ",
"Cok ",
"Tuan ",
"Xiu ",
"Guan ",
"Sen ",
"Lyen ",
"Shou ",
"O ",
"Man ",
"Mak ",
"Luo ",
"Bi ",
"Wei ",
"Liu ",
"Cek ",
"Qiao ",
"Cong ",
"Yi ",
"Lu ",
"O ",
"Kayng ",
"Cang ",
"Cui ",
"Qi ",
"Chang ",
"Tang ",
"Man ",
"Yong ",
"San ",
"Feng ",
"Kyeng ",
"Phyo ",
"Shu ",
"Lwu ",
"Xiu ",
"Chong ",
"Long ",
"Cham ",
"Jian ",
"Cao ",
"Li ",
"Xia ",
"Xi ",
"Kang ",
"",
"Beng ",
"",
"",
"Zheng ",
"Lu ",
"Hua ",
"Cip ",
"Pu ",
"Hui ",
"Qiang ",
"Po ",
"Lin ",
"Suo ",
"Swu ",
"San ",
"Cheng ",
],
"x94": [
"Kui ",
"Si ",
"Liu ",
"Nyo ",
"Hoyng ",
"Pie ",
"Sui ",
"Fan ",
"Qiao ",
"Quan ",
"Yang ",
"Tang ",
"Xiang ",
"Jue ",
"Jiao ",
"Cwun ",
"Lyo ",
"Jie ",
"Lao ",
"Tay ",
"Sim ",
"Zan ",
"Ji ",
"Jian ",
"Cong ",
"Tung ",
"Ya ",
"Ying ",
"Dui ",
"Jue ",
"Nou ",
"Ti ",
"Pu ",
"Tie ",
"",
"",
"Ding ",
"Sen ",
"Kai ",
"Jian ",
"Fei ",
"Sui ",
"Lo ",
"Cen ",
"Hui ",
"Yu ",
"Lian ",
"Zhuo ",
"Qiao ",
"Qian ",
"Zhuo ",
"Lei ",
"Bi ",
"Chel ",
"Hwan ",
"Ye ",
"Thak ",
"Guo ",
"Tang ",
"Ju ",
"Fen ",
"Da ",
"Bei ",
"Yi ",
"Ai ",
"Zong ",
"Hwun ",
"Diao ",
"Cwu ",
"Heng ",
"Zhui ",
"Ji ",
"Nie ",
"Ta ",
"Hwak ",
"Qing ",
"Pin ",
"Ying ",
"Kui ",
"Ning ",
"Xu ",
"Kam ",
"Kam ",
"Yari ",
"Cha ",
"Cil ",
"Mie ",
"Li ",
"Lei ",
"Ji ",
"Zuan ",
"Kwang ",
"Shang ",
"Peng ",
"Lap ",
"Du ",
"Sak ",
"Chuo ",
"Lye ",
"Phyo ",
"Bao ",
"Lu ",
"",
"Thoa ",
"Long ",
"E ",
"Lo ",
"Xin ",
"Jian ",
"Lan ",
"Bo ",
"Jian ",
"Yak ",
"Cham ",
"Yang ",
"Jian ",
"Xi ",
"Kwan ",
"Cang ",
"Sep ",
"Lei ",
"Cuan ",
"Qu ",
"Pan ",
"La ",
"Chan ",
"Lan ",
"Chak ",
"Nie ",
"Jue ",
"Tang ",
"Shu ",
"Lan ",
"Jin ",
"Qiu ",
"Yi ",
"Zhen ",
"Ding ",
"Zhao ",
"Po ",
"Diao ",
"Tu ",
"Qian ",
"Chuan ",
"Shan ",
"Sap ",
"Fan ",
"Diao ",
"Men ",
"Nu ",
"Xi ",
"Chai ",
"Xing ",
"Gai ",
"Bu ",
"Tai ",
"Ju ",
"Dun ",
"Chao ",
"Zhong ",
"Na ",
"Bei ",
"Gang ",
"Ban ",
"Qian ",
"Yao ",
"Qin ",
"Jun ",
"Wu ",
"Gou ",
"Kang ",
"Fang ",
"Huo ",
"Dou ",
"Niu ",
"Ba ",
"Yu ",
"Qian ",
"Zheng ",
"Qian ",
"Gu ",
"Bo ",
"E ",
"Po ",
"Bu ",
"Ba ",
"Yue ",
"Zuan ",
"Mu ",
"Dan ",
"Jia ",
"Dian ",
"You ",
"Tie ",
"Bo ",
"Ling ",
"Shuo ",
"Qian ",
"Liu ",
"Bao ",
"Shi ",
"Xuan ",
"She ",
"Bi ",
"Ni ",
"Pi ",
"Duo ",
"Xing ",
"Kao ",
"Lao ",
"Er ",
"Mang ",
"Ya ",
"You ",
"Cheng ",
"Jia ",
"Ye ",
"Nao ",
"Zhi ",
"Dang ",
"Tong ",
"Lu ",
"Diao ",
"Yin ",
"Kai ",
"Zha ",
"Zhu ",
"Xian ",
"Ting ",
"Diu ",
"Xian ",
"Hua ",
"Quan ",
"Sha ",
"Jia ",
"Yao ",
"Ge ",
"Ming ",
"Zheng ",
"Se ",
"Jiao ",
"Yi ",
"Chan ",
"Chong ",
"Tang ",
"An ",
"Yin ",
"Ru ",
"Zhu ",
"Lao ",
"Pu ",
"Wu ",
"Lai ",
"Te ",
"Lian ",
"Keng ",
],
"x95": [
"Xiao ",
"Suo ",
"Li ",
"Zheng ",
"Chu ",
"Guo ",
"Gao ",
"Tie ",
"Xiu ",
"Cuo ",
"Lue ",
"Feng ",
"Xin ",
"Liu ",
"Kai ",
"Jian ",
"Rui ",
"Ti ",
"Lang ",
"Qian ",
"Ju ",
"A ",
"Qiang ",
"Duo ",
"Tian ",
"Cuo ",
"Mao ",
"Ben ",
"Ki ",
"De ",
"Kua ",
"Kun ",
"Chang ",
"Xi ",
"Gu ",
"Luo ",
"Chui ",
"Zhui ",
"Jin ",
"Zhi ",
"Xian ",
"Juan ",
"Huo ",
"Pou ",
"Tan ",
"Ding ",
"Jian ",
"Ju ",
"Meng ",
"Zi ",
"Qie ",
"Yeng ",
"Kai ",
"Qiang ",
"Song ",
"E ",
"Cha ",
"Qiao ",
"Cong ",
"Duan ",
"Sou ",
"Koyng ",
"Huan ",
"Ai ",
"Du ",
"Mei ",
"Lou ",
"Zi ",
"Fei ",
"Mei ",
"Mo ",
"Zhen ",
"Bo ",
"Ge ",
"Nie ",
"Tang ",
"Juan ",
"Nie ",
"Na ",
"Liu ",
"Hao ",
"Bang ",
"Yi ",
"Jia ",
"Bin ",
"Rong ",
"Biao ",
"Tang ",
"Man ",
"Luo ",
"Beng ",
"Yong ",
"Jing ",
"Di ",
"Zu ",
"Xuan ",
"Liu ",
"Tan ",
"Jue ",
"Liao ",
"Pu ",
"Lu ",
"Dui ",
"Lan ",
"Pu ",
"Cuan ",
"Qiang ",
"Deng ",
"Huo ",
"Lei ",
"Huan ",
"Zhuo ",
"Lian ",
"Yi ",
"Cha ",
"Biao ",
"La ",
"Chan ",
"Xiang ",
"Cang ",
"Chang ",
"Jiu ",
"Ao ",
"Die ",
"Qu ",
"Liao ",
"Mi ",
"Chang ",
"Mwun ",
"Ma ",
"Shuan ",
"Sem ",
"Huo ",
"Men ",
"Yan ",
"Bi ",
"Han ",
"Phyey ",
"San ",
"Kay ",
"Kang ",
"Beng ",
"Koyng ",
"Lyun ",
"San ",
"Han ",
"Han ",
"Kan ",
"Min ",
"Xia ",
"Yuru ",
"Dou ",
"Kap ",
"Nao ",
"",
"Peng ",
"Xia ",
"Ling ",
"Bian ",
"Pi ",
"Run ",
"He ",
"Kwan ",
"Kak ",
"Hap ",
"Pel ",
"Chwuk ",
"Hong ",
"Kyu ",
"Min ",
"Se ",
"Kon ",
"Lang ",
"Lye ",
"Ting ",
"Sha ",
"Ju ",
"Yel ",
"Yel ",
"Chan ",
"Qu ",
"Lin ",
"Cheng ",
"Shai ",
"Kun ",
"Em ",
"Wen ",
"Yem ",
"An ",
"Hon ",
"Yek ",
"Wen ",
"Xiang ",
"Bao ",
"Xiang ",
"Kyek ",
"Yao ",
"Wen ",
"Ban ",
"Am ",
"Wi ",
"Yin ",
"Hwal ",
"Kyel ",
"Lan ",
"To ",
"",
"Phwung ",
"Tian ",
"Nie ",
"Ta ",
"Kay ",
"Hap ",
"Kwel ",
"Chum ",
"Kwan ",
"Dou ",
"Qi ",
"Kyu ",
"Tang ",
"Kwan ",
"Piao ",
"Ham ",
"Xi ",
"Kwey ",
"Chen ",
"Pyek ",
"Dang ",
"Huan ",
"Tal ",
"Wen ",
"",
"Men ",
"Shuan ",
"Shan ",
"Yan ",
"Han ",
"Bi ",
"Wen ",
"Chuang ",
"Run ",
"Wei ",
"Xian ",
"Hong ",
"Jian ",
"Min ",
"Kang ",
"Men ",
"Zha ",
"Nao ",
"Gui ",
"Wen ",
"Ta ",
"Min ",
"Lu ",
"Kai ",
],
"x96": [
"Fa ",
"Ge ",
"He ",
"Kun ",
"Jiu ",
"Yue ",
"Lang ",
"Du ",
"Yu ",
"Yan ",
"Chang ",
"Xi ",
"Wen ",
"Hun ",
"Yan ",
"E ",
"Chan ",
"Lan ",
"Qu ",
"Hui ",
"Kuo ",
"Que ",
"Ge ",
"Tian ",
"Ta ",
"Que ",
"Kan ",
"Huan ",
"Pwu ",
"Pwu ",
"Le ",
"Dui ",
"Xin ",
"Chen ",
"Ol ",
"Yi ",
"Chi ",
"Yin ",
"Yang ",
"Dou ",
"Ayk ",
"Sheng ",
"Phan ",
"Pei ",
"Keng ",
"Yun ",
"Wan ",
"Ci ",
"Pi ",
"Ceng ",
"Pang ",
"Yang ",
"Yin ",
"Zhen ",
"Jie ",
"Cheng ",
"E ",
"Qu ",
"Di ",
"Co ",
"Co ",
"Cem ",
"Ling ",
"A ",
"Tha ",
"Tuo ",
"Phi ",
"Bing ",
"Pwu ",
"Ji ",
"Lu ",
"Long ",
"Chen ",
"Xing ",
"Duo ",
"Lwu ",
"Mayk ",
"Kang ",
"Shu ",
"La ",
"Han ",
"Er ",
"Gui ",
"Yu ",
"Hay ",
"Shan ",
"Xun ",
"Qiao ",
"Hyeng ",
"Chun ",
"Fu ",
"Phyey ",
"Hyep ",
"Sem ",
"Sung ",
"Chek ",
"Pu ",
"Twu ",
"Wen ",
"Cin ",
"Cey ",
"Ham ",
"Tou ",
"Nie ",
"Yun ",
"Xian ",
"Pay ",
"Pei ",
"Chwu ",
"Yi ",
"Dui ",
"Lun ",
"Um ",
"Ju ",
"Swu ",
"Cin ",
"Pi ",
"Lung ",
"To ",
"Ham ",
"Lyuk ",
"",
"Hem ",
"Yin ",
"Ce ",
"Yang ",
"Reng ",
"Shan ",
"Chong ",
"Yan ",
"Yin ",
"Yu ",
"Cey ",
"Wu ",
"Lyung ",
"Wei ",
"Oy ",
"El ",
"Tay ",
"Swu ",
"An ",
"Hwang ",
"Kyey ",
"Swu ",
"Un ",
"Ki ",
"Yan ",
"Hui ",
"Kyek ",
"Wun ",
"Wu ",
"Oy ",
"Ay ",
"Kuk ",
"Tang ",
"Cey ",
"Cang ",
"Dao ",
"Ao ",
"Xi ",
"Un ",
"",
"Rao ",
"Lin ",
"Thoy ",
"Deng ",
"Pi ",
"Swu ",
"Swu ",
"O ",
"Hem ",
"Fen ",
"Ni ",
"Er ",
"Ji ",
"Dao ",
"Sup ",
"Un ",
"E ",
"Hyu ",
"Long ",
"Xi ",
"I ",
"Lyey ",
"Lyey ",
"Chwu ",
"He ",
"Chek ",
"Cwun ",
"Jun ",
"Nan ",
"Yi ",
"Cak ",
"An ",
"Qin ",
"Ya ",
"Wung ",
"A ",
"Cip ",
"Ko ",
"Huan ",
"Chi ",
"Gou ",
"Cwun ",
"Ca ",
"Ong ",
"Ce ",
"Chu ",
"Hu ",
"Cap ",
"Lak ",
"Yu ",
"Chou ",
"Co ",
"Swu ",
"Han ",
"Huo ",
"Ssang ",
"Kwan ",
"Chwu ",
"Cap ",
"Ong ",
"Kyey ",
"Xi ",
"Chou ",
"Liu ",
"Li ",
"Nan ",
"Xue ",
"Za ",
"Ji ",
"Ji ",
"Wu ",
"Wu ",
"Sel ",
"Na ",
"Fou ",
"Se ",
"Mu ",
"Mwun ",
"Pwun ",
"Pang ",
"Wun ",
"Li ",
"Li ",
"Ang ",
"Lyeng ",
"Loy ",
"An ",
"Pak ",
"Mong ",
"Cen ",
"Dang ",
"Xing ",
"Wu ",
"Zhao ",
],
"x97": [
"Swu ",
"Ji ",
"Mu ",
"Chen ",
"So ",
"Sap ",
"Ceng ",
"Cin ",
"Phay ",
"May ",
"Ling ",
"Qi ",
"Cwu ",
"Kwak ",
"Sap ",
"Pi ",
"Weng ",
"Cem ",
"Yin ",
"Yey ",
"Cwu ",
"Tun ",
"Lim ",
"",
"Dong ",
"Yeng ",
"Wu ",
"Ling ",
"Sang ",
"Ling ",
"Ha ",
"Hong ",
"Yin ",
"Mo ",
"Mai ",
"Wun ",
"Liu ",
"Meng ",
"Pin ",
"Mwu ",
"Wei ",
"Huo ",
"Um ",
"Xi ",
"Yi ",
"Ai ",
"Dan ",
"Deng ",
"Sen ",
"Yu ",
"Lo ",
"Long ",
"Dai ",
"Ji ",
"Pang ",
"Yang ",
"Phay ",
"Pyek ",
"Wei ",
"",
"Xi ",
"Cey ",
"May ",
"Meng ",
"Meng ",
"Lei ",
"Lyek ",
"Huo ",
"Ay ",
"Fei ",
"Chey ",
"Long ",
"Lyeng ",
"Ay ",
"Feng ",
"Li ",
"Po ",
"",
"He ",
"He ",
"Bing ",
"Cheng ",
"Cheng ",
"Jing ",
"Tian ",
"Zhen ",
"Ceng ",
"Cheng ",
"Qing ",
"Jing ",
"Ceng ",
"Cen ",
"Ceng ",
"Chen ",
"Pi ",
"Fei ",
"Ko ",
"Mi ",
"Myen ",
"Mian ",
"Pao ",
"Ye ",
"Cen ",
"Hui ",
"Yep ",
"Hyek ",
"Ding ",
"Cha ",
"Jian ",
"In ",
"Di ",
"Du ",
"Wu ",
"Ren ",
"Qin ",
"Kun ",
"Hwa ",
"Nyu ",
"Pha ",
"In ",
"Sa ",
"Na ",
"Mal ",
"Zu ",
"Tal ",
"Ban ",
"Yi ",
"Yao ",
"To ",
"Phi ",
"Jia ",
"Hong ",
"Pho ",
"Ang ",
"Tomo ",
"Yin ",
"Jia ",
"Tao ",
"Ji ",
"Hyey ",
"An ",
"An ",
"Hen ",
"Kong ",
"Kohaze ",
"Da ",
"Qiao ",
"Ting ",
"Man ",
"Ying ",
"Sui ",
"Tiao ",
"Cho ",
"Xuan ",
"Kong ",
"Beng ",
"Ta ",
"Zhang ",
"Bing ",
"Kuo ",
"Kwuk ",
"La ",
"Xie ",
"Yu ",
"Bang ",
"Yi ",
"Chwu ",
"Qiu ",
"Kal ",
"Xiao ",
"Mu ",
"Kwu ",
"Ken ",
"Phyen ",
"Cey ",
"Jian ",
"On ",
"To ",
"Kwu ",
"Ta ",
"Pi ",
"Xie ",
"Pan ",
"Ge ",
"Phil ",
"Kwak ",
"Tou ",
"Lou ",
"Gui ",
"Kyo ",
"Xue ",
"Ji ",
"Jian ",
"Kang ",
"Chan ",
"Tal ",
"Huo ",
"Xian ",
"Chen ",
"Du ",
"Wa ",
"Chen ",
"Lan ",
"Wi ",
"Ren ",
"Pwul ",
"Mei ",
"Juan ",
"Kap ",
"Wei ",
"Qiao ",
"Han ",
"Chang ",
"",
"Rou ",
"Xun ",
"She ",
"Wi ",
"Ge ",
"Bei ",
"To ",
"Kwu ",
"On ",
"",
"Phil ",
"Wi ",
"Hui ",
"Du ",
"Wa ",
"Du ",
"Wei ",
"Ren ",
"Fu ",
"Han ",
"Wei ",
"Yun ",
"Tao ",
"Kwu ",
"Kwu ",
"Xian ",
"Xie ",
"Sem ",
"Cey ",
"Um ",
"Za ",
"Wun ",
"So ",
"Le ",
"Peng ",
"Heng ",
"Yeng ",
"Wun ",
"Peng ",
"Yin ",
"Yin ",
"Hyang ",
],
"x98": [
"Ho ",
"Hyel ",
"Ceng ",
"Kyeng ",
"Kui ",
"Hang ",
"Swun ",
"Han ",
"Swu ",
"I ",
"Wuk ",
"Gu ",
"Song ",
"Kyu ",
"Ki ",
"Hang ",
"Yey ",
"Wan ",
"Pan ",
"Ton ",
"Di ",
"Dan ",
"Pan ",
"Pha ",
"Lyeng ",
"Ce ",
"Kyeng ",
"Lei ",
"He ",
"Qiao ",
"Al ",
"E ",
"Wei ",
"Hil ",
"Gua ",
"Sin ",
"I ",
"Shen ",
"Hay ",
"Dui ",
"Pian ",
"Ping ",
"Lei ",
"Pwu ",
"Hyep ",
"Twu ",
"Hoy ",
"Kui ",
"Hyep ",
"Le ",
"Ting ",
"Cheng ",
"Ying ",
"Jun ",
"Hu ",
"Am ",
"Kyeng ",
"Thoy ",
"Tui ",
"Pin ",
"Loy ",
"Thoy ",
"Zi ",
"Ca ",
"Chui ",
"Ding ",
"Loy ",
"Yan ",
"Han ",
"Jian ",
"Kwa ",
"Chwi ",
"Kyeng ",
"Qin ",
"Yi ",
"Si ",
"Cey ",
"Ayk ",
"Ak ",
"An ",
"Hun ",
"Kan ",
"Ong ",
"Cen ",
"An ",
"Hyen ",
"Sin ",
"Yi ",
"Wen ",
"Sang ",
"Cen ",
"Cen ",
"Jiang ",
"Ku ",
"Lyu ",
"Liao ",
"Piao ",
"Yi ",
"Man ",
"Qi ",
"Rao ",
"Ho ",
"Cho ",
"Ko ",
"Xun ",
"Qian ",
"Hui ",
"Cen ",
"Ru ",
"Hong ",
"Bin ",
"Hyen ",
"Pin ",
"Lo ",
"Lan ",
"Sep ",
"Kwan ",
"Ye ",
"Ding ",
"Qing ",
"Han ",
"Xiang ",
"Shun ",
"Xu ",
"Xu ",
"Wan ",
"Gu ",
"Dun ",
"Qi ",
"Ban ",
"Song ",
"Hang ",
"Yu ",
"Lu ",
"Ling ",
"Po ",
"Jing ",
"Jie ",
"Jia ",
"Tian ",
"Han ",
"Ying ",
"Jiong ",
"Hai ",
"Yi ",
"Pin ",
"Hui ",
"Tui ",
"Han ",
"Ying ",
"Ying ",
"Ke ",
"Ti ",
"Yong ",
"E ",
"Zhuan ",
"Yan ",
"E ",
"Nie ",
"Man ",
"Dian ",
"Sang ",
"Hao ",
"Lei ",
"Zhan ",
"Ru ",
"Pin ",
"Quan ",
"Phwung ",
"Biao ",
"Oroshi ",
"Fu ",
"Xia ",
"Cem ",
"Biao ",
"Sap ",
"Ba ",
"Thay ",
"Lyel ",
"Kwal ",
"Xuan ",
"Shao ",
"Kwu ",
"Bi ",
"Si ",
"Wei ",
"Yang ",
"Yao ",
"Swu ",
"Kai ",
"Sao ",
"Pem ",
"Liu ",
"Xi ",
"Lyo ",
"Piao ",
"Phyo ",
"Liu ",
"Phyo ",
"Phyo ",
"Biao ",
"Lyo ",
"",
"Sil ",
"Feng ",
"Biao ",
"Feng ",
"Yang ",
"Zhan ",
"Biao ",
"Sa ",
"Ju ",
"Si ",
"Sou ",
"Yao ",
"Liu ",
"Piao ",
"Biao ",
"Biao ",
"Pi ",
"Pen ",
"Fei ",
"Fei ",
"Sik ",
"Shi ",
"Son ",
"Ki ",
"Ceng ",
"Si ",
"Tuo ",
"Cen ",
"Son ",
"Xiang ",
"Tun ",
"Im ",
"E ",
"Juan ",
"Chik ",
"Um ",
"Pan ",
"Fan ",
"Son ",
"Um ",
"Zhu ",
"I ",
"Zhai ",
"Bi ",
"Jie ",
"Tao ",
"Liu ",
"Ci ",
"Chel ",
"Sa ",
"Pho ",
"Sik ",
"Duo ",
],
"x99": [
"Hai ",
"Im ",
"Tian ",
"Kyo ",
"Jia ",
"Pyeng ",
"Yao ",
"Tong ",
"Ci ",
"Hyang ",
"Yang ",
"Yang ",
"I ",
"Yan ",
"Le ",
"Yi ",
"Chan ",
"Bo ",
"Noy ",
"A ",
"Pho ",
"Cwun ",
"Dou ",
"Su ",
"Ye ",
"Shi ",
"Hyo ",
"Hun ",
"Guo ",
"Shi ",
"Cen ",
"Zhui ",
"Pyeng ",
"Xian ",
"Bu ",
"Ye ",
"Tan ",
"Fei ",
"Cang ",
"Wi ",
"Kwan ",
"E ",
"Nan ",
"Hun ",
"Ho ",
"Huang ",
"Chel ",
"Hui ",
"Cen ",
"Hwu ",
"He ",
"Tang ",
"Fen ",
"Wei ",
"Gu ",
"Cha ",
"Song ",
"Tang ",
"Bo ",
"Gao ",
"Huy ",
"Kwey ",
"Liu ",
"Sou ",
"Tao ",
"Ye ",
"On ",
"Mo ",
"Tang ",
"Man ",
"Bi ",
"Yu ",
"Swu ",
"Kun ",
"Can ",
"Kwey ",
"Chan ",
"Sen ",
"Chi ",
"Dan ",
"Uy ",
"Ki ",
"Yo ",
"Cheng ",
"Ong ",
"To ",
"Hui ",
"Hyang ",
"Zhan ",
"Fen ",
"Hai ",
"Meng ",
"Yem ",
"Mo ",
"Cham ",
"Xiang ",
"Luo ",
"Zuan ",
"Nang ",
"Shi ",
"Ceng ",
"Ji ",
"Tuo ",
"Xing ",
"Tun ",
"Xi ",
"Ren ",
"Yu ",
"Chi ",
"Fan ",
"Yin ",
"Jian ",
"Shi ",
"Bao ",
"Si ",
"Duo ",
"Yi ",
"Er ",
"Rao ",
"Xiang ",
"Jia ",
"Le ",
"Jiao ",
"Yi ",
"Bing ",
"Bo ",
"Dou ",
"E ",
"Yu ",
"Nei ",
"Jun ",
"Guo ",
"Hun ",
"Xian ",
"Guan ",
"Cha ",
"Kui ",
"Gu ",
"Sou ",
"Chan ",
"Ye ",
"Mo ",
"Bo ",
"Liu ",
"Xiu ",
"Jin ",
"Man ",
"Can ",
"Zhuan ",
"Nang ",
"Swu ",
"Kyu ",
"Koyk ",
"Hyang ",
"Fen ",
"Ba ",
"Ni ",
"Phil ",
"Bo ",
"Tu ",
"Han ",
"Fei ",
"Jian ",
"Am ",
"Ai ",
"Pok ",
"Xian ",
"Wen ",
"Hyeng ",
"Fen ",
"Bin ",
"Xing ",
"Ma ",
"E ",
"Phwung ",
"Han ",
"Cek ",
"Tha ",
"Tuo ",
"Chi ",
"Swun ",
"Zhu ",
"Zhi ",
"Pei ",
"Xin ",
"Il ",
"Sa ",
"Yin ",
"Wen ",
"Zhi ",
"Dan ",
"Lu ",
"You ",
"Pak ",
"Bao ",
"Kuai ",
"Tha ",
"Yek ",
"Kwu ",
"",
"Kwu ",
"Kyeng ",
"Bo ",
"Zhao ",
"Yuan ",
"Peng ",
"Zhou ",
"Ke ",
"Cwu ",
"No ",
"Kwu ",
"Pi ",
"Zang ",
"Ka ",
"Ling ",
"Zhen ",
"Tha ",
"Pwu ",
"Yang ",
"Sa ",
"Phil ",
"Tha ",
"Tha ",
"Sa ",
"Liu ",
"Ma ",
"Pyen ",
"Tao ",
"Zhi ",
"Rong ",
"Teng ",
"Dong ",
"Swun ",
"Quan ",
"Sin ",
"Jiong ",
"Er ",
"Hay ",
"Pak ",
"",
"Yin ",
"Lak ",
"Shuu ",
"Dan ",
"Xie ",
"Liu ",
"Ju ",
"Song ",
"Chim ",
"Mang ",
"Liang ",
"Han ",
"Tu ",
"Hyen ",
"Tui ",
"Cwun ",
],
"x9a": [
"E ",
"Ping ",
"Seng ",
"Ay ",
"Lok ",
"Chwu ",
"Zhou ",
"She ",
"Pyeng ",
"Kun ",
"Tao ",
"Lay ",
"Zong ",
"Kwa ",
"Ki ",
"Ki ",
"Yan ",
"Pi ",
"So ",
"Hem ",
"Jie ",
"Yao ",
"Mwu ",
"Pian ",
"Chong ",
"Phyen ",
"Qian ",
"Pi ",
"Huang ",
"Jian ",
"Huo ",
"Yu ",
"Ti ",
"Quan ",
"Xia ",
"Zong ",
"Kui ",
"Rou ",
"Si ",
"Gua ",
"Tuo ",
"Kui ",
"Sou ",
"Ken ",
"Cheng ",
"Cul ",
"Liu ",
"Pang ",
"Tung ",
"Xi ",
"Cao ",
"Du ",
"Yan ",
"Wen ",
"Chwu ",
"So ",
"Sen ",
"Li ",
"Zhi ",
"Shuang ",
"Lu ",
"Xi ",
"La ",
"Zhang ",
"Mayk ",
"O ",
"Cham ",
"Phyo ",
"Chong ",
"Kwu ",
"Bi ",
"Zhi ",
"Yu ",
"Xu ",
"Hwa ",
"Bo ",
"Swuk ",
"Hyo ",
"Lin ",
"Can ",
"Dun ",
"Liu ",
"Than ",
"Zeng ",
"Tan ",
"Kyo ",
"Tie ",
"Hem ",
"La ",
"Zhan ",
"Kyeng ",
"Yek ",
"Ye ",
"Tuo ",
"Bin ",
"Chwi ",
"Yan ",
"Peng ",
"Lye ",
"Teng ",
"Yang ",
"Ki ",
"Shuang ",
"Ju ",
"Xi ",
"Hwan ",
"Lye ",
"Biao ",
"Ma ",
"Yu ",
"Tuo ",
"Xun ",
"Chi ",
"Qu ",
"Il ",
"Bo ",
"Lu ",
"Zang ",
"Shi ",
"Si ",
"Fu ",
"Ju ",
"Zou ",
"Zhu ",
"Tuo ",
"Nu ",
"Jia ",
"Yi ",
"Tai ",
"Xiao ",
"Ma ",
"Yin ",
"Jiao ",
"Hua ",
"Luo ",
"Hai ",
"Pian ",
"Biao ",
"Li ",
"Cheng ",
"Yan ",
"Xin ",
"Qin ",
"Jun ",
"Qi ",
"Qi ",
"Ke ",
"Zhui ",
"Zong ",
"Swuk ",
"Can ",
"Pian ",
"Zhi ",
"Kui ",
"Sao ",
"Wu ",
"Ao ",
"Liu ",
"Qian ",
"Shan ",
"Piao ",
"Luo ",
"Cong ",
"Chan ",
"Zou ",
"Ji ",
"Shuang ",
"Xiang ",
"Kol ",
"Wei ",
"I ",
"I ",
"Yu ",
"Gan ",
"Yi ",
"Hang ",
"Thwu ",
"Xie ",
"Bao ",
"Bi ",
"Chi ",
"Ti ",
"Cey ",
"Ko ",
"Hay ",
"Kyo ",
"Gou ",
"Kua ",
"Kyek ",
"Tui ",
"Geng ",
"Pyen ",
"Pi ",
"Kwa ",
"Ka ",
"Yu ",
"Sui ",
"Lou ",
"Pak ",
"Xiao ",
"Pang ",
"Bo ",
"Ci ",
"Kuan ",
"Bin ",
"Mo ",
"Liao ",
"Lwu ",
"Nao ",
"Chok ",
"Cang ",
"Swu ",
"Chey ",
"Pin ",
"Kwan ",
"Lo ",
"Ko ",
"Ko ",
"Qiao ",
"Kao ",
"Qiao ",
"Lao ",
"Zao ",
"Phyo ",
"Kun ",
"Kon ",
"Ti ",
"Pang ",
"Xiu ",
"Yem ",
"Mo ",
"Dan ",
"Kun ",
"Bin ",
"Fa ",
"Cho ",
"Pi ",
"Ca ",
"Pal ",
"Yem ",
"Chey ",
"Pao ",
"Phi ",
"Mao ",
"Pwul ",
"Er ",
"Rong ",
"Qu ",
"",
"Hyu ",
"Kwal ",
"Kyey ",
"Peng ",
"Cwa ",
"Shao ",
"Sha ",
],
"x9b": [
"Ti ",
"Li ",
"Pin ",
"Zong ",
"Ti ",
"Pwul ",
"Song ",
"Zheng ",
"Kwen ",
"Zong ",
"Swun ",
"Cen ",
"Duo ",
"Ho ",
"La ",
"Jiu ",
"Ki ",
"Lian ",
"Cin ",
"Bin ",
"Peng ",
"Mo ",
"Sam ",
"Man ",
"Man ",
"Sung ",
"Swu ",
"Lie ",
"Qian ",
"Qian ",
"Nong ",
"Hwan ",
"Kwal ",
"Ning ",
"Pin ",
"Lyep ",
"Rang ",
"Dou ",
"Dou ",
"Lyo ",
"Hong ",
"Hyek ",
"Thwu ",
"Han ",
"Dou ",
"Dou ",
"Kwu ",
"Cheng ",
"Wul ",
"Wul ",
"Kyek ",
"Juan ",
"Fu ",
"Qian ",
"Gui ",
"Zong ",
"Liu ",
"Gui ",
"Sang ",
"Yu ",
"Kwi ",
"Mei ",
"Ji ",
"Qi ",
"Jie ",
"Koy ",
"Hon ",
"Pal ",
"Payk ",
"May ",
"Xu ",
"Yan ",
"So ",
"Liang ",
"Yu ",
"Chwu ",
"Qi ",
"Mang ",
"Lyang ",
"Wi ",
"Jian ",
"Li ",
"Piao ",
"Bi ",
"Ma ",
"Ji ",
"Xu ",
"Chou ",
"Yem ",
"Zhan ",
"E ",
"Dao ",
"In ",
"Ji ",
"Eri ",
"Gong ",
"Tuo ",
"Diao ",
"Ji ",
"Xu ",
"E ",
"E ",
"Sa ",
"Hang ",
"Tun ",
"Mo ",
"Jie ",
"Shen ",
"Fan ",
"Yuan ",
"Bi ",
"Lo ",
"Wen ",
"Hu ",
"Lu ",
"Za ",
"Pang ",
"Fen ",
"Na ",
"You ",
"Namazu ",
"Todo ",
"He ",
"Xia ",
"Qu ",
"Han ",
"Pi ",
"Ling ",
"Tha ",
"Bo ",
"Qiu ",
"Phyeng ",
"Fu ",
"Bi ",
"Ji ",
"Wei ",
"Ju ",
"Diao ",
"Bo ",
"You ",
"Gun ",
"Pi ",
"Cem ",
"Xing ",
"Thay ",
"Pho ",
"Pwu ",
"Ca ",
"Ju ",
"Gu ",
"Kajika ",
"Tong ",
"",
"Ta ",
"Kil ",
"Shu ",
"Hou ",
"Sang ",
"I ",
"An ",
"Wei ",
"Tiao ",
"Zhu ",
"Yin ",
"Lie ",
"Luo ",
"Tong ",
"Cey ",
"Ci ",
"Bing ",
"Yu ",
"Kyo ",
"Bu ",
"Hay ",
"Sen ",
"Ge ",
"Hui ",
"Bora ",
"Mate ",
"Kao ",
"Gori ",
"Duo ",
"Jun ",
"Ti ",
"Mian ",
"So ",
"Za ",
"Sha ",
"Qin ",
"Yu ",
"Nei ",
"Zhe ",
"Kon ",
"Kyeng ",
"",
"Wu ",
"Qiu ",
"Ting ",
"Pho ",
"Hon ",
"Tiao ",
"Li ",
"Sa ",
"Sa ",
"Gao ",
"Meng ",
"Ugui ",
"Asari ",
"Subashiri ",
"Kazunoko ",
"Yong ",
"Ni ",
"Chi ",
"Qi ",
"Cheng ",
"Xiang ",
"Nei ",
"Chun ",
"Ji ",
"Co ",
"Qie ",
"Ko ",
"Zhou ",
"Dong ",
"Lai ",
"Pi ",
"Yey ",
"Yi ",
"Kon ",
"Lu ",
"Jiu ",
"Chang ",
"Kyeng ",
"Lun ",
"Lung ",
"Chwu ",
"Li ",
"Meng ",
"Zong ",
"Zhi ",
"Nian ",
"Shachi ",
"Dojou ",
"Sukesou ",
"Shi ",
"Shen ",
"Hun ",
"Cey ",
"Hou ",
"Xing ",
"Zhu ",
"La ",
"Zong ",
"Cuk ",
"Bian ",
"Phyen ",
],
"x9c": [
"Huan ",
"Quan ",
"Ze ",
"Wei ",
"Wei ",
"Yu ",
"Qun ",
"Rou ",
"Cep ",
"Hwang ",
"Lyen ",
"Yan ",
"Chwu ",
"Chwu ",
"Jian ",
"Bi ",
"Ak ",
"Yang ",
"Pok ",
"Say ",
"Jian ",
"Ha ",
"Tuo ",
"Hu ",
"Muroaji ",
"Ruo ",
"Haraka ",
"Wen ",
"Jian ",
"Hao ",
"Wu ",
"Fang ",
"Sao ",
"Liu ",
"Ma ",
"Si ",
"Sa ",
"Hwan ",
"Shi ",
"Teng ",
"Thap ",
"Yo ",
"Ge ",
"Rong ",
"Qian ",
"Ki ",
"On ",
"Yak ",
"Hatahata ",
"Lyen ",
"O ",
"Luk ",
"Hui ",
"Min ",
"Ji ",
"Co ",
"Qu ",
"Kyen ",
"So ",
"Man ",
"Xi ",
"Qiu ",
"Phyo ",
"Ji ",
"Ji ",
"Zhu ",
"Jiang ",
"Qiu ",
"Zhuan ",
"Yong ",
"Zhang ",
"Kang ",
"Sel ",
"Pyel ",
"Jue ",
"Qu ",
"Xiang ",
"Bo ",
"Jiao ",
"Sim ",
"Su ",
"Huang ",
"Cwun ",
"Sen ",
"Sen ",
"Fan ",
"Kwel ",
"Lin ",
"Sim ",
"Miao ",
"Xi ",
"Eso ",
"Kyou ",
"Pwun ",
"Guan ",
"Hwu ",
"Hoy ",
"Zei ",
"Sao ",
"Cen ",
"Gan ",
"Gui ",
"Sheng ",
"Lyey ",
"Chang ",
"Hatahata ",
"Shiira ",
"Ai ",
"Ru ",
"Cey ",
"Xu ",
"Huo ",
"Shiira ",
"Li ",
"Lie ",
"Li ",
"Mie ",
"Zhen ",
"Xiang ",
"E ",
"Lo ",
"Guan ",
"Lye ",
"Sen ",
"Yu ",
"Dao ",
"Ji ",
"You ",
"Tun ",
"Lu ",
"Fang ",
"Ba ",
"He ",
"Bo ",
"Ping ",
"Nian ",
"Lu ",
"You ",
"Zha ",
"Fu ",
"Bo ",
"Bao ",
"Hou ",
"Pi ",
"Tai ",
"Gui ",
"Jie ",
"Kao ",
"Wei ",
"Er ",
"Tong ",
"Ze ",
"Hou ",
"Kuai ",
"Ji ",
"Jiao ",
"Xian ",
"Za ",
"Xiang ",
"Xun ",
"Geng ",
"Li ",
"Lian ",
"Jian ",
"Li ",
"Shi ",
"Tiao ",
"Gun ",
"Sha ",
"Wan ",
"Jun ",
"Ji ",
"Yong ",
"Qing ",
"Ling ",
"Qi ",
"Zou ",
"Fei ",
"Kun ",
"Chang ",
"Gu ",
"Ni ",
"Nian ",
"Diao ",
"Jing ",
"Shen ",
"Shi ",
"Zi ",
"Fen ",
"Die ",
"Bi ",
"Chang ",
"Shi ",
"Wen ",
"Wei ",
"Sai ",
"E ",
"Qiu ",
"Fu ",
"Huang ",
"Quan ",
"Jiang ",
"Bian ",
"Sao ",
"Ao ",
"Qi ",
"Ta ",
"Yin ",
"Yao ",
"Fang ",
"Jian ",
"Le ",
"Biao ",
"Xue ",
"Bie ",
"Man ",
"Min ",
"Yong ",
"Wei ",
"Xi ",
"Jue ",
"Shan ",
"Lin ",
"Zun ",
"Huo ",
"Gan ",
"Li ",
"Zhan ",
"Guan ",
"Co ",
"Ul ",
"Pwu ",
"Li ",
"Kwu ",
"Bu ",
"Yan ",
"Pwu ",
"Diao ",
"Kyey ",
"Pong ",
"Nio ",
"Gan ",
"Shi ",
"Pong ",
"Myeng ",
"Bao ",
"Yen ",
"Zhi ",
"Hu ",
"Qin ",
"Fu ",
"Fen ",
"Wen ",
"Jian ",
"Shi ",
"Yu ",
],
"x9d": [
"Fou ",
"Yao ",
"Jue ",
"Kyek ",
"Pi ",
"Huan ",
"Cim ",
"Po ",
"An ",
"A ",
"Zheng ",
"Fang ",
"Pong ",
"Wen ",
"Ou ",
"Te ",
"Jia ",
"Nu ",
"Lyeng ",
"Mie ",
"Fu ",
"Tha ",
"Wen ",
"Li ",
"Pyen ",
"Chi ",
"Ge ",
"Wen ",
"Zi ",
"Qu ",
"Xiao ",
"Chi ",
"Dan ",
"Ju ",
"You ",
"Ko ",
"Zhong ",
"Yu ",
"Ang ",
"Rong ",
"Ap ",
"Tie ",
"Yu ",
"Shigi ",
"Ying ",
"Zhui ",
"Wu ",
"Er ",
"Kwal ",
"Ai ",
"Zhi ",
"Yan ",
"Heng ",
"Jiao ",
"Al ",
"Lie ",
"Zhu ",
"Ren ",
"Yi ",
"Hong ",
"Luo ",
"Ru ",
"Mou ",
"Ge ",
"Ren ",
"Kyo ",
"Hyu ",
"Zhou ",
"Chi ",
"Luo ",
"Chidori ",
"Toki ",
"Ten ",
"Luan ",
"Jia ",
"Ji ",
"Yu ",
"Huan ",
"Tuo ",
"Bu ",
"Wu ",
"Kyen ",
"Yu ",
"Pal ",
"Cwun ",
"Xun ",
"Bi ",
"Xi ",
"Jun ",
"Ju ",
"Tu ",
"Jing ",
"Cey ",
"A ",
"A ",
"Kuang ",
"Kok ",
"Mwu ",
"Shen ",
"Lai ",
"Ikaruga ",
"Kakesu ",
"Lu ",
"Ping ",
"Shu ",
"Pok ",
"An ",
"Zhao ",
"Pwung ",
"Qin ",
"Qian ",
"Phil ",
"Co ",
"Lu ",
"Cak ",
"Jian ",
"Ju ",
"Tu ",
"Ya ",
"Wen ",
"Qi ",
"Li ",
"Ye ",
"Chwu ",
"Kong ",
"Zhui ",
"Kon ",
"Sheng ",
"Qi ",
"Jing ",
"Yi ",
"Yi ",
"Cheng ",
"Zi ",
"Lai ",
"Dong ",
"Qi ",
"Swun ",
"Geng ",
"Ju ",
"Qu ",
"Isuka ",
"Kikuitadaki ",
"Ji ",
"Shu ",
"",
"Chi ",
"Miao ",
"Rou ",
"An ",
"Chwu ",
"Ti ",
"Hu ",
"Ti ",
"Ak ",
"Jie ",
"Mao ",
"Fu ",
"Chun ",
"Tu ",
"Yan ",
"Kal ",
"Yuan ",
"Pian ",
"Kon ",
"Mei ",
"Hu ",
"Ying ",
"Dun ",
"Mok ",
"Ju ",
"Tsugumi ",
"Cheng ",
"Fang ",
"Gu ",
"Ayng ",
"Yuan ",
"Xuan ",
"Weng ",
"Shi ",
"Hak ",
"Chwu ",
"Tang ",
"Xia ",
"Yak ",
"Lyu ",
"Ji ",
"Kol ",
"Jian ",
"Zhun ",
"Han ",
"Ca ",
"Zi ",
"Ik ",
"Yo ",
"Yan ",
"Kyey ",
"Li ",
"Tian ",
"Kwu ",
"Ti ",
"Ti ",
"Ni ",
"Tu ",
"Ma ",
"Jiao ",
"Gao ",
"Tian ",
"Chen ",
"Li ",
"Zhuan ",
"Ca ",
"Ao ",
"Yao ",
"Yey ",
"Kwu ",
"Chi ",
"Ci ",
"Liao ",
"Rong ",
"Lou ",
"Bi ",
"Shuang ",
"Zhuo ",
"Yu ",
"Wu ",
"Jue ",
"Yin ",
"Quan ",
"Si ",
"Cho ",
"Yi ",
"Hua ",
"Pyel ",
"Ayng ",
"Swuk ",
"Huang ",
"Fan ",
"Jiao ",
"Liao ",
"Yen ",
"Ko ",
"Chwi ",
"Han ",
"Xian ",
"Tu ",
"Mai ",
"Zun ",
"Hul ",
"Ung ",
"Lo ",
"Tuan ",
"Xian ",
"Hak ",
"Yi ",
"Pyek ",
],
"x9e": [
"Shu ",
"Luo ",
"Qi ",
"Yi ",
"Ji ",
"Zhe ",
"Yu ",
"Cen ",
"Ye ",
"Yang ",
"Pi ",
"Ning ",
"Huo ",
"Mi ",
"Ying ",
"Mong ",
"Di ",
"Yue ",
"Yu ",
"Lei ",
"Bao ",
"Lo ",
"He ",
"Long ",
"Shuang ",
"Yue ",
"Ayng ",
"Kwan ",
"Kwu ",
"Li ",
"Lan ",
"Niao ",
"Jiu ",
"Ji ",
"Yuan ",
"Ming ",
"Shi ",
"Ou ",
"Ya ",
"Cang ",
"Bao ",
"Zhen ",
"Gu ",
"Dong ",
"Lu ",
"Ya ",
"Xiao ",
"Yang ",
"Ling ",
"Zhi ",
"Qu ",
"Yuan ",
"Xue ",
"Tuo ",
"Si ",
"Zhi ",
"Er ",
"Gua ",
"Xiu ",
"Heng ",
"Zhou ",
"Ge ",
"Luan ",
"Hong ",
"Wu ",
"Bo ",
"Li ",
"Juan ",
"Hu ",
"E ",
"Yu ",
"Xian ",
"Ti ",
"Wu ",
"Que ",
"Miao ",
"An ",
"Kun ",
"Bei ",
"Peng ",
"Qian ",
"Chun ",
"Geng ",
"Yuan ",
"Su ",
"Hu ",
"He ",
"E ",
"Gu ",
"Chwu ",
"Zi ",
"Mei ",
"Mu ",
"Ni ",
"Yao ",
"Weng ",
"Liu ",
"Ji ",
"Ni ",
"Jian ",
"He ",
"Yi ",
"Ying ",
"Zhe ",
"Liao ",
"Liao ",
"Jiao ",
"Jiu ",
"Yu ",
"Lu ",
"Xuan ",
"Zhan ",
"Ying ",
"Huo ",
"Meng ",
"Guan ",
"Shuang ",
"Lo ",
"Kung ",
"Lyeng ",
"Jian ",
"Ham ",
"Cuo ",
"Kem ",
"Jian ",
"Yem ",
"Cuo ",
"Lok ",
"Wu ",
"Cu ",
"Kwey ",
"Biao ",
"Chwu ",
"Biao ",
"Zhu ",
"Kyun ",
"Cwu ",
"Jian ",
"Mi ",
"Mi ",
"Wu ",
"Liu ",
"Chen ",
"Kyun ",
"Lin ",
"Yey ",
"Ki ",
"Lok ",
"Jiu ",
"Jun ",
"Jing ",
"Li ",
"Xiang ",
"Yan ",
"Jia ",
"Mi ",
"Li ",
"Sa ",
"Cang ",
"Lin ",
"Jing ",
"Ji ",
"Ling ",
"Yan ",
"Chwu ",
"Mayk ",
"Mayk ",
"Ge ",
"Chao ",
"Pwu ",
"Myen ",
"Mian ",
"Fu ",
"Pao ",
"Qu ",
"Kwuk ",
"Mo ",
"Fu ",
"Xian ",
"Lai ",
"Kwuk ",
"Myen ",
"",
"Feng ",
"Fu ",
"Kwuk ",
"Myen ",
"Ma ",
"Ma ",
"Ma ",
"Hwi ",
"Ma ",
"Zou ",
"Nen ",
"Fen ",
"Hwang ",
"Hwang ",
"Kum ",
"Guang ",
"Tian ",
"Tou ",
"Heng ",
"Xi ",
"Kuang ",
"Hoyng ",
"Se ",
"Lye ",
"Cem ",
"Li ",
"Huk ",
"Huk ",
"Yi ",
"Kem ",
"Dan ",
"Ki ",
"Tuan ",
"Mwuk ",
"Mo ",
"Kyem ",
"Tay ",
"Chwul ",
"Yu ",
"Cem ",
"I ",
"Hil ",
"Yan ",
"Qu ",
"Mei ",
"Yan ",
"Kyeng ",
"Yu ",
"Li ",
"Tang ",
"Du ",
"Can ",
"Yin ",
"An ",
"Yan ",
"Tam ",
"Am ",
"Zhen ",
"Dai ",
"Cham ",
"Yi ",
"Mi ",
"Tam ",
"Yem ",
"Tok ",
"Lu ",
"Zhi ",
"Pwun ",
"Pwul ",
"Po ",
"Min ",
"Min ",
"Wen ",
],
"x9f": [
"Chwuk ",
"Qu ",
"Co ",
"Wa ",
"Cwu ",
"Zhi ",
"Mang ",
"O ",
"Pyel ",
"Tha ",
"Pyek ",
"Yuan ",
"Chao ",
"Tuo ",
"Ceng ",
"Mi ",
"Nay ",
"Ding ",
"Zi ",
"Ko ",
"Gu ",
"Tong ",
"Fen ",
"To ",
"Yuan ",
"Pi ",
"Chang ",
"Gao ",
"Qi ",
"Yuan ",
"Tang ",
"Teng ",
"Se ",
"Shu ",
"Pwun ",
"Fei ",
"Wen ",
"Ba ",
"Diao ",
"Tuo ",
"Tong ",
"Qu ",
"Sheng ",
"Sek ",
"Yu ",
"Shi ",
"Ting ",
"O ",
"Nian ",
"Jing ",
"Hun ",
"Ju ",
"En ",
"Tu ",
"Ti ",
"Hyey ",
"Kyem ",
"En ",
"Lei ",
"Pi ",
"Yao ",
"Kwu ",
"Han ",
"Wu ",
"Wu ",
"Hou ",
"Xi ",
"Ge ",
"Cha ",
"Xiu ",
"Weng ",
"Zha ",
"Nong ",
"Nang ",
"Ca ",
"Cay ",
"Ji ",
"Zi ",
"Cay ",
"Cey ",
"Qi ",
"Ji ",
"Chi ",
"Chen ",
"Chin ",
"Hul ",
"Ya ",
"Un ",
"Xie ",
"Pho ",
"Cuo ",
"Shi ",
"Cha ",
"Chi ",
"Nian ",
"Ce ",
"Cho ",
"Lyeng ",
"Ling ",
"Chek ",
"Quan ",
"Xie ",
"Kan ",
"Sel ",
"Jiu ",
"Kyo ",
"Chak ",
"Kon ",
"E ",
"Chu ",
"Yi ",
"Ni ",
"Cuo ",
"Zou ",
"Wu ",
"Nen ",
"Xian ",
"Ou ",
"Ak ",
"Ak ",
"Yi ",
"Chuo ",
"Zou ",
"Dian ",
"Chu ",
"Jin ",
"Ya ",
"Chi ",
"Chen ",
"He ",
"Ken ",
"Ju ",
"Ling ",
"Pao ",
"Tiao ",
"Zi ",
"Ken ",
"Yu ",
"Chuo ",
"Qu ",
"Wo ",
"Lyong ",
"Pang ",
"Gong ",
"Pang ",
"Yan ",
"Lyong ",
"Long ",
"Kong ",
"Kam ",
"Ta ",
"Ling ",
"Ta ",
"Long ",
"Gong ",
"Kan ",
"Kwu ",
"Chwu ",
"Pyel ",
"Gui ",
"Yak ",
"Chwi ",
"Hwa ",
"Jue ",
"Hay ",
"Yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Shan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"xf9": [
"Kay ",
"Kayng ",
"Ke ",
"Ko ",
"Kol ",
"Koc ",
"Kwi ",
"Kwi ",
"Kyun ",
"Kul ",
"Kum ",
"Na ",
"Na ",
"Na ",
"Na ",
"Na ",
"Na ",
"Na ",
"Na ",
"Na ",
"Nak ",
"Nak ",
"Nak ",
"Nak ",
"Nak ",
"Nak ",
"Nak ",
"Nan ",
"Nan ",
"Nan ",
"Nan ",
"Nan ",
"Nan ",
"Nam ",
"Nam ",
"Nam ",
"Nam ",
"Nap ",
"Nap ",
"Nap ",
"Nang ",
"Nang ",
"Nang ",
"Nang ",
"Nang ",
"Nay ",
"Nayng ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"No ",
"Nok ",
"Nok ",
"Nok ",
"Nok ",
"Nok ",
"Nok ",
"Non ",
"Nong ",
"Nong ",
"Nong ",
"Nong ",
"Noy ",
"Noy ",
"Noy ",
"Noy ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nwu ",
"Nuk ",
"Nuk ",
"Num ",
"Nung ",
"Nung ",
"Nung ",
"Nung ",
"Nung ",
"Twu ",
"La ",
"Lak ",
"Lak ",
"Lan ",
"Lyeng ",
"Lo ",
"Lyul ",
"Li ",
"Pey ",
"Pen ",
"Pyen ",
"Pwu ",
"Pwul ",
"Pi ",
"Sak ",
"Sak ",
"Sam ",
"Sayk ",
"Sayng ",
"Sep ",
"Sey ",
"Sway ",
"Sin ",
"Sim ",
"Sip ",
"Ya ",
"Yak ",
"Yak ",
"Yang ",
"Yang ",
"Yang ",
"Yang ",
"Yang ",
"Yang ",
"Yang ",
"Yang ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Ye ",
"Yek ",
"Yek ",
"Yek ",
"Yek ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yen ",
"Yel ",
"Yel ",
"Yel ",
"Yel ",
"Yel ",
"Yel ",
"Yem ",
"Yem ",
"Yem ",
"Yem ",
"Yem ",
"Yep ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yeng ",
"Yey ",
"Yey ",
"Yey ",
"Yey ",
"O ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yo ",
"Yong ",
"Wun ",
"Wen ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yu ",
"Yuk ",
"Yuk ",
"Yuk ",
"Yun ",
"Yun ",
"Yun ",
"Yun ",
"Yul ",
"Yul ",
"Yul ",
"Yul ",
"Yung ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"I ",
"Ik ",
"Ik ",
"In ",
"In ",
"In ",
"In ",
"In ",
"In ",
"In ",
"Im ",
"Im ",
"Im ",
"Ip ",
"Ip ",
"Ip ",
"Cang ",
"Cek ",
"Ci ",
"Cip ",
"Cha ",
"Chek ",
],
"xfa": [
"Chey ",
"Thak ",
"Thak ",
"Thang ",
"Thayk ",
"Thong ",
"Pho ",
"Phok ",
"Hang ",
"Hang ",
"Hyen ",
"Hwak ",
"Ol ",
"Huo ",
"",
"Coc ",
"Chong ",
"",
"Cheng ",
"",
"",
"Huy ",
"Ce ",
"Ik ",
"Lyey ",
"Sin ",
"Sang ",
"Pok ",
"Ceng ",
"Ceng ",
"Wu ",
"",
"Hagi ",
"",
"Cey ",
"",
"",
"Il ",
"To ",
"",
"",
"",
"Pan ",
"Sa ",
"Kwan ",
"Hak ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x200": [
"Ha ",
"Qi ",
"",
"",
"",
"Hai ",
"",
"",
"",
"Qiu ",
"",
"",
"",
"Shi ",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ba ",
"Cup ",
"",
"Kha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trut ",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Duoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"Cui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ga ",
"",
"",
"",
"",
"Nham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Suot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zi ",
"",
"",
"",
"",
"",
"",
"",
"",
"Zung ",
"",
"",
"",
"",
"Pai ",
"",
"Dui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
],
"x201": [
"Vu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gop ",
"Vai ",
"",
"Hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kep ",
"",
"",
"Nham ",
"",
"",
"",
"",
"Lam ",
"Nam ",
"Vai ",
"",
"wei ",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mat ",
"Mat ",
"",
"Mat ",
"",
"Mat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhang ",
"",
"",
"Mat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Yong ",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Faan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trum ",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Voi ",
"",
"",
"",
"",
"",
"",
"",
"Va ",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hua ",
"Top ",
"",
"",
],
"x202": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nay ",
"",
"Tray ",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"Ngai ",
"",
"Nho ",
"Thay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bing ",
"",
"zhuan ",
"shu ",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"she ",
"",
"",
"",
"",
"",
"",
"May ",
"Mu ",
"",
"",
"fu ",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tao ",
"",
"",
"",
"",
"",
],
"x203": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Wu ",
"",
"",
"",
"",
"",
"",
"",
"shuai ",
"",
"",
"gai ",
"",
"",
"",
"zhuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"Man ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"But ",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"Gap ",
"",
"bie ",
"",
"",
"qu ",
"",
"",
"yang ",
"",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xum ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cap ",
"Bay ",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"sa ",
"",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"yu ",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngai ",
"",
"",
"",
"",
"",
"",
"",
"Tui ",
"",
"",
"",
"",
"",
"Giong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x204": [
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"shu ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trom ",
"",
"",
"Long ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngua ",
"",
"",
"",
"",
"nei ",
"nei ",
"",
"",
"zhai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhau ",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"Tin ",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"Truoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quanh ",
"",
"",
"",
"Mong ",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vao ",
"Nhui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lon ",
"",
"Tron ",
"Sip ",
"",
"",
"",
],
"x205": [
"",
"",
"",
"Xi ",
"",
"",
"",
"",
"",
"Juan ",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"lun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"heng ",
"",
"",
"",
"",
"",
"",
"",
"",
"Zheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"qia ",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xuong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bay ",
"Chong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"Lun ",
"",
"",
"Ping ",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"Tron ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ret ",
"",
"",
"",
"liu ",
"",
"",
],
"x206": [
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lai ",
"",
"",
"",
"",
"he ",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"chang ",
"",
"",
"",
"",
"",
"Moc ",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ra ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Got ",
"",
"zhe ",
"",
"",
"",
"ju ",
"",
"",
"",
"Shan ",
"Sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chem ",
"",
"",
"",
"",
"",
"Tiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"",
"",
"",
"",
],
"x207": [
"",
"",
"",
"jiu ",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xian ",
"",
"kuai ",
"",
"",
"",
"",
"",
"",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hui ",
"",
"",
"Pai ",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"Qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cun ",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gaai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Got ",
"Bam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gun ",
"Lou ",
"",
"",
"Jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hoat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ,bai",
"",
"",
"",
"",
],
"x208": [
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"Bua ",
"",
"",
"",
"",
"",
"",
"",
"Tet ",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"Truoc ",
"",
"",
"",
"Chom ",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"duan ",
"",
"",
"",
"",
"",
"",
"",
"",
"Khuot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"Vam ",
"Giup ",
"Giup ",
"",
"",
"",
"dian ",
"Jue ",
"",
"",
"",
"",
"",
"",
"",
"Lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ruon ",
"",
"",
"",
"",
"",
"",
"10383.060,wan ",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"shao ",
"",
"",
"",
"",
"",
"",
"",
"pao ",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"fu ",
"jiu ",
],
"x209": [
"",
"",
"",
"",
"",
"",
"Cho ",
"",
"",
"",
"",
"",
"Cho ",
"",
"hua ",
"Bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"luan ",
"Muoi ",
"",
"",
"",
"",
"",
"Gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chuc ",
"",
"Lung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tron ",
"yu ",
"",
"",
"",
"",
"",
"Nhu ",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"shao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"heng ",
],
"x20a": [
"",
"",
"",
"wai ",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"zhuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"bang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Me ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Li ",
"",
"",
"",
"",
"",
"",
"",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"Day ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"",
],
"x20b": [
"",
"",
"",
"",
"",
"",
"guan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cut ",
"",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"",
"fu ",
"",
"",
"Mot ",
"",
"",
"bang ",
"",
"",
"",
"",
"",
"",
"biao ",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"Jin ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zhuo ",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"de ",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Naai ",
"Leo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mang ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngot ",
"sa ,san",
"",
"",
"",
"",
"",
"",
"",
"",
"Daai6",
"",
"",
"Jai ",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"ban ",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thet ",
"",
"",
"",
"",
"",
"",
"",
"Hin ",
"pian ",
"",
"",
"",
"",
"bian ",
"",
"",
"reng ",
"",
"reng ",
"",
"",
"Danh ",
"Chui ",
"",
"",
],
"x20c": [
"",
"",
"Ngoen ",
"",
"",
"",
"",
"",
"",
"",
"",
"Jaau ",
"Mut ",
"",
"",
"Mom ",
"",
"an ",
"",
"he ",
"",
"",
"",
"",
"",
"Khen ",
"",
"hu ,gao",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Meo ",
"",
"",
"",
"",
"Eot ",
"",
"",
"",
"",
"Lo ",
"",
"",
"dai ",
"wai ,wai",
"",
"",
"",
"",
"",
"Tam ",
"Dat ",
"Nip ",
"",
"Quai ",
"",
"",
"",
"",
"",
"",
"",
"Phom ",
"",
"",
"",
"",
"",
"Ngai ",
"",
"",
"",
"",
"Ngaak6",
"",
"",
"",
"Chun ",
"",
"",
"",
"",
"sa ,shai",
"",
"",
"",
"Fik ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"na ",
"",
"",
"",
"",
"Ming ",
"San ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"Nham ",
"",
"",
"",
"Hang ",
"",
"",
"E ",
"",
"",
"",
"",
"",
"Gianh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhung ",
"",
"",
"",
"",
"",
"",
"",
"Khinh ",
"",
"",
"ge ",
"",
"",
"",
"Mep ",
"",
"die ",
"",
"",
"",
"",
"",
"fu ",
"",
"Shu ",
"",
"Kwan ",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"qia ",
"",
"",
"Ce ",
"Vang ",
"",
"",
"",
"",
"Caau ",
"Dap6",
"",
"Nhu ",
"Thay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tu ",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
],
"x20d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Phao ",
"Nhanh ",
"Nhan ",
"",
"Mang ",
"Nuc ",
"",
"",
"",
"",
"Miu ",
"Voi ",
"",
"",
"",
"",
"Gung ",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"Ngoam ",
"zhu ",
"",
"",
"",
"",
"Thut ",
"",
"",
"",
"",
"Gau6",
"hu ",
"ta ",
"Ngaau ",
"",
"zao ",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"Daam ",
"Koe ",
"Mui ",
"Hong ",
"",
"",
"",
"",
"",
"",
"Mep ",
"",
"",
"",
"",
"",
"",
"",
"Mun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"zhen ,chun",
"de ",
"Go ",
"",
"",
"",
"",
"",
"Gwit ",
"",
"E ",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"Mang ",
"Faat ",
"Waak ",
"Pe ",
"Tham ",
"Nhu ",
"",
"",
"",
"",
"",
"",
"Se ",
"",
"Pha ",
"",
"",
"",
"",
"",
"",
"U ",
"",
"",
"",
"Nhau ",
"Uong ",
"",
"Sat ",
"Bop ",
"",
"",
"Zai ",
"",
"",
"",
"",
"Troi ",
"Du ",
"Mai ",
"",
"Khung ",
"",
"Dim6",
"",
"",
"",
"da ",
"nang ",
"",
"",
"Chut ",
"",
"",
"Gap6",
"",
"",
"",
"",
"",
"jue ",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Phao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x20e": [
"",
"pai ",
"Giot ",
"",
"Inh ",
"",
"",
"",
"",
"Kak ",
"Tap ",
"",
"",
"Me ",
"Naa ",
"Ge ",
"Kam ",
"Soek ",
"Bou ",
"",
"",
"",
"",
"Xua ",
"Tuc ",
"Song ",
"",
"",
"",
"",
"",
"Bai ",
"",
"",
"",
"",
"",
"",
"Khan ",
"",
"",
"",
"",
"Tau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"Ngaak6",
"",
"",
"",
"Map ",
"",
"xun ",
"wa ",
"",
"ang ",
"han ",
"",
"",
"",
"",
"",
"",
"",
"Lan ",
"",
"",
"",
"",
"yao ",
"",
"ge ",
"",
"bao ",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ko ",
"",
"",
"",
"",
"",
"Git6",
"",
"",
"",
"Ngo ",
"Kam ",
"Lan ",
"Maai ",
"Zam6",
"",
"Cay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gwe ",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"Long ",
"Thoi ",
"Ban ",
"",
"",
"Gaak ",
"",
"ku ",
"Lung ",
"",
"Gaa ",
"",
"",
"",
"",
"",
"",
"Trou ",
"He ",
"Loe ",
"He ",
"",
"",
"",
"",
"Hung ",
"",
"",
"",
"Chac ",
"Nop ",
"",
"",
"Ri ",
"Que ",
"Cop ",
"Xui ",
"",
"Chau ",
"Ngoan ",
"",
"Guong ",
"Ngon ",
"Han ",
"Oang ",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"Le ",
"Zeot6",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Don ",
"zhao ",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"Long ",
"",
"",
"",
"",
"",
"Aa6",
"Bai ",
"Dau6",
"",
"",
"",
],
"x20f": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Paai ",
"",
"Zaam ",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiang ,qiang",
"",
"",
"",
"Muon ",
"",
"",
"lun ",
"Day ",
"",
"",
"But ",
"Ngai ",
"Ho ",
"Kang ",
"Loe ",
"",
"",
"",
"Danh ",
"",
"Thay ",
"",
"",
"",
"Ji ",
"",
"",
"",
"",
"",
"Xo ",
"",
"Zap ",
"Tham ",
"Thung ",
"Nuot ",
"",
"",
"",
"",
"Nac ",
"Toet ",
"",
"Nhai ",
"",
"Ngo ",
"",
"Neng ",
"Ngon ",
"Thua ",
"",
"Giuc ",
"",
"",
"",
"",
"Oam ",
"",
"",
"",
"",
"",
"",
"",
"",
"Kik ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"Hup ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"He ",
"",
"Ce ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngoang ",
"",
"",
"",
"",
"",
"",
"shu ",
"Rum ",
"",
"",
"Bai ",
"",
"",
"la ",
"",
"",
"",
"We ",
"",
"",
"",
"",
"",
"",
"Baang ",
"Zaa ",
"Ging ",
"",
"",
"Nuot ",
"",
"",
"Cyut ",
"Nhun ",
"Nhap ",
"",
"",
"Si ",
"Xep ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Het ",
"sa ",
"",
"qiao ",
"",
"",
"",
"",
"",
"Lu ",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gaa ",
"Saau ",
"Soe ",
"Wet ",
"",
"Ngui ",
"",
"",
"",
"",
"",
"Khan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x210": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nam6",
"Zip ",
"",
"Bei6",
"",
"",
"",
"",
"Phao ",
"",
"",
"",
"Lok ",
"",
"Bam ",
"",
"",
"",
"",
"",
"Hao ",
"Cay ",
"",
"",
"",
"",
"Ron ",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"ai ",
"yu ",
"Am ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hon ",
"",
"Wo ",
"Hang ",
"Xao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dyut ",
"",
"",
"",
"zhan ",
"",
"Gan ",
"Zit ",
"Doeng ",
"Kwaat ",
"",
"Ngon ",
"Ziu ",
"",
"",
"",
"",
"Khao ",
"",
"",
"Hun ",
"",
"",
"",
"Mom ",
"Voc ",
"",
"yu ",
"",
"",
"Eng ",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"Lai ",
"",
"",
"",
"",
"",
"",
"Zuk6",
"",
"bao ",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ze ",
"peng ",
"",
"",
"Ngau ",
"",
"Ran ",
"",
"Ray ",
"Troi ",
"",
"",
"",
"Khoan ",
"",
"",
"",
"",
"",
"",
"",
"",
"E ",
"Leng ",
"",
"",
"lai ",
"",
"",
"Bai6",
"Kwaak ",
"Gaa ",
"",
"",
"",
"Chem ",
"Phan ",
"Doe6",
"",
"",
"Boc ",
"Bo ",
"",
"Gung ",
"Le ",
"Mua ",
"",
"Mut ",
"",
"",
"",
"lun ",
"",
"",
"",
"",
"",
"",
"Laai6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ceoi ",
"Ngung ",
"Sek ",
"",
"",
"Chen ",
"",
"",
"",
"",
"Phac ",
"Thot ",
],
"x211": [
"",
"Lum ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ruc ",
"Gam ",
"Ham ",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Haa ",
"",
"Chay ",
"",
"",
"",
"",
"Lom ",
"",
"",
"Khan ",
"Toe ",
"Khem ",
"Hun ",
"",
"",
"Jik ",
"Tot ",
"cu ",
"ru ",
"",
"",
"Ji ",
"",
"",
"",
"luan ",
"",
"",
"",
"Soe ",
"",
"",
"",
"",
"",
"",
"",
"e ,ei",
"",
"",
"Laa ",
"Dang ",
"Bun ",
"Hum ",
"",
"",
"",
"Lai ",
"Lanh ",
"Ngong ",
"",
"",
"",
"",
"",
"",
"la ",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"Nan ",
"",
"",
"",
"",
"Hoai ",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"Gwang ",
"Nhau ",
"",
"",
"Nhep ",
"wan ",
"",
"Wok ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nham ",
"",
"",
"",
"lian ",
"Trom ",
"",
"",
"",
"Gu ",
"",
"",
"",
"Giau ",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"Doe ",
"",
"",
"",
"",
"he ",
"",
"zhe ",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"",
"dang ",
"he ",
"tai ",
"Quay ",
"",
"yu ",
"Nhot ",
"ya ",
"",
"",
"",
],
"x212": [
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"Mui ",
"",
"",
"",
"",
"yuan ",
"",
"",
"",
"",
"",
"You ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cong ",
"",
"",
"",
"tun ",
"",
"",
"",
"Cong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ,hu",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"Tum ",
"Moc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"de ",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x213": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"guai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"Che ",
"Vung ",
"Lam ",
"",
"Mun ",
"Nui ",
"",
"",
"",
"",
"",
"",
"",
"zhai ",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"an ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bun ",
"Nam ",
"",
"",
"Hang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thong ",
"su ",
"",
"",
"",
"",
"",
"Gung ",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"",
"Luong ",
"",
"",
"",
"Tret ",
"Xay ",
"Bui ",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x214": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"Hoc ",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ve ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Coi ",
"Chum ",
"",
"",
"",
"Thoai ",
"",
"",
"",
"",
"",
"Lan ",
"Sui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Coi ",
"",
"",
"",
"",
"Gom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"sa ",
"",
"Juk ",
"Lan ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"ju ",
"",
"",
"",
"",
"shu ",
"xian ",
"",
"",
"gai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cau ",
"",
"",
"Suong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Loi ",
"Vung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"San ",
"",
"",
"lai ",
"",
"Lam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chen ",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"Chum ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x215": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kua ",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"guai ",
"",
"",
"",
"",
"su ",
"",
"Bon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sai ",
"",
"",
"",
"",
"",
"",
"",
"",
"Lam ",
"",
"",
"Tum ",
"",
"Muong ",
"",
"",
"",
"",
"",
"",
"Peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"",
"",
],
"x216": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"Mat ",
"Lon ",
"juan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Canh ",
"Nhon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bi ",
"",
"Xon ",
"",
"",
"",
"yan ",
"wei ",
"",
"",
"",
"hao ",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"Lon ",
"",
"",
"",
"",
"",
"jue ",
"To ",
"To ",
"",
"",
"",
"Hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhang ",
"",
"da ",
"",
"hao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"Nua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x217": [
"fan ",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"Ji ",
"gao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bua ",
"",
"",
"",
"",
"chu ",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kep ",
"Va ",
"",
"",
"",
"",
"cheng ",
"",
"du ",
"",
"",
"nian ",
"",
"",
"",
"",
"",
"",
"Vu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"pan ",
"hu ",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"fu ",
"nai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x218": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"yang ",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"Cuoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"Nhang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuan ",
"",
"",
"",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"",
"Sau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"man ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nei ",
"",
"",
"",
"",
"",
"Cuoi ",
"",
"",
"",
"",
"",
"cai ",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sen ",
"",
"",
"",
"",
],
"x219": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chua ",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"Moi ",
"Moi ",
"",
"",
"zhu ",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"Nhat ",
"xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Con ",
"ju ,ru",
"",
"",
"",
"",
"",
"Me ",
"",
"",
"",
"",
"xu ",
"",
"",
"fu ",
"So ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"er ",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rot ",
"",
"",
"",
"Chat ",
"",
"",
"Nhang ",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mian ",
"",
"",
"",
"",
"wa ",
"",
"pao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"heng ",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
],
"x21a": [
"",
"",
"",
"",
"Xia ",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"Tot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"Ning ",
"nai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"an ",
"",
"",
"Xum ",
"",
"",
"",
"ceng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x21b": [
"",
"",
"",
"",
"",
"",
"",
"",
"Xuong ",
"sai ",
"yu ",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"20960.020,lao ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tac ",
"",
"",
"",
"",
"Gang ",
"",
"",
"",
"",
"",
"de ",
"",
"",
"",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"shuan ",
"",
"",
"",
"",
"Cut ",
"",
"",
"",
"",
"",
"",
"Mon ",
"",
"",
"",
"",
"",
"bai ",
"",
"",
"",
"",
"Chut ",
"",
"",
"",
"",
"Be ",
"",
"",
"",
"",
"Chut ",
"Trut ",
"kao ",
"",
"",
"",
"",
"luan ",
"",
"",
"Nhon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mon ",
"Chut ",
"",
"Mon ",
"",
"May ",
"Be ",
"Chut ",
"",
"",
"",
"",
"",
"Choai ",
"",
"",
"",
"",
"",
"nai ",
"",
"",
"",
"",
"",
"",
"May ",
"",
"",
"Be ",
"Be ",
"Be ",
"Zao ",
"",
"",
"Be ",
"",
"Nhen ",
"Mon ",
"Nhon ",
"Mon ",
"",
"",
"",
"Tho ",
"",
"",
"Chuong ",
"Chuong ",
"",
"Nhon ",
"",
"",
"",
"",
"Nhon ",
"",
"",
"Oat ",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ga ",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"",
"",
"",
"ga ",
"",
"",
],
"x21c": [
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"Duk ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"Shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"Thuoc ",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"die ",
"Gang ",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ti ",
"",
"jue ",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"Ke ",
"jiang ",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"Gwat6",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngoe ",
"",
"kai ",
"Cuoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"Cuoi ",
"Rot ",
"",
"",
"",
"",
"",
"",
"",
"Toi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cuoi ",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
],
"x21d": [
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"hu ,jie",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngut ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"ang ",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cu ",
"",
"",
"",
"",
"",
"Nui ",
"",
"yao ",
"ai ",
"",
"",
"",
"",
"",
"fan ",
"",
"ju ",
"",
"",
"",
"",
"qie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuang ",
"",
"",
"",
"ya ",
"",
"Ngan ",
"",
"kan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"han ",
"",
"na ",
"",
"",
"",
"",
"",
"",
"wu ",
"gao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
],
"x21e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"jie ",
"Voi ",
"",
"",
"",
"",
"xie ",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nghi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Din ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Voi ",
"",
"Seoi ",
"Von ",
"Chon ",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"gun ",
"man ",
"",
"",
"zu ",
"",
"hu ",
"",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x21f": [
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuai ",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Non ",
"",
"",
"",
"",
"",
"Jie ",
"bang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ceng ",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"Doc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nang ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nao ",
"",
"xun ",
"",
"ju ",
"",
],
"x220": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vang ",
"Sua ",
"Tron ",
"Sang ",
"",
"",
"",
"Lon ",
"",
"Sam ",
"",
"",
"To ",
"Lon ",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trang ",
"",
"Va ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Phuon ",
"",
"",
"bu ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"xuan ",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"Du ",
"",
"xie ",
"Shi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tranh ",
"yuan ",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"qian ",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"Lei ",
"Long ",
"",
"Vua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ta ",
"",
"han ",
"qian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x221": [
"",
"",
"",
"Phoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ,ge",
"bang ",
"",
"meng ",
"",
"",
"wu ",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"han ",
"Bau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"Man ",
"",
"",
"",
"",
"",
"xiang ",
"Hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Manh ",
"",
"",
"mao ",
"",
"ceng ",
"",
"",
"",
"Lian ",
"",
"",
"dan ",
"",
"",
"",
"Lian ",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"gai ",
"Ju ",
"",
"",
"",
"",
"zu ",
"",
"Chan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"",
"nei ",
"",
"",
"",
"",
"",
"neng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"Jian ",
"",
"Lai ",
"Nin ",
"",
"",
"",
"Nam ",
"jian ",
"May ",
"May ",
"",
"May ",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Guan ",
"yan ",
"Doi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"Khuya ",
"",
"",
"",
"",
"ren ",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bu ",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x222": [
"",
"",
"",
"",
"",
"",
"",
"",
"Nap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"",
"",
"Roku ",
"",
"",
"kun ",
"",
"qiu ",
"",
"",
"cu ,la",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ,tu",
"",
"",
"",
"",
"",
"hu ",
"",
"jia ",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kho ",
"U ",
"ma ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
],
"x223": [
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"Rong ",
"",
"Rong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gao ",
"juan ",
"",
"",
"",
"",
"",
"quan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"dai ",
"",
"su ",
"",
"",
"",
"",
"",
"jie ",
"",
"qu ",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"juan ",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"jue ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cong ",
"Cong ",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"Beng ",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"Von ",
],
"x224": [
"",
"",
"",
"",
"yuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"King ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bay ",
"",
"",
"",
"fu ",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ang ",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"kua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"wu ,hu",
"",
"",
"xiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x225": [
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"liu ",
"",
"20832.030,yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shuai ",
"",
"",
"",
"",
"",
"yuan ",
"",
"",
"",
"",
"",
"shuai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"sa ",
"xian ",
"zha ",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"San ",
"",
"",
"shan ",
"",
"",
"",
"",
"Suot ",
"",
"",
"",
"",
"",
"",
"",
"tiao ",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"Sau ",
"Sau ",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"jue ",
"nang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"Nuoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"chang ",
"",
"Lo ",
"",
"",
"",
],
"x226": [
"",
"",
"",
"",
"qiu ",
"Maau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rung ",
"Rap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hung ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhon ",
"E ",
"Tim ",
"kan ",
"",
"",
"Lung6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"miao ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"Ti ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khuay ",
"",
"zhu ",
"Ngop ",
"",
"",
"",
"",
"",
"",
"qiu ",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"ru ",
"wang ",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mang ",
"Zang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x227": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xing ",
"Duo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"Nao ",
"",
"",
"",
"",
"",
"",
"jia ",
"tu ",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"Fit ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gao ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"Fit ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chua ",
"",
"Nan ",
"",
"",
"Nep ",
"",
"",
"Chac ",
"Cham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"teng ",
"",
"",
"",
"",
"",
"",
"",
"Nhuoc ",
"",
"",
"",
"",
"",
"",
],
"x228": [
"",
"",
"",
"Geng6",
"Sung ",
"Thung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"",
"Gan ",
"Hon ",
"",
"",
"Mo ",
"",
"shu ",
"",
"",
"",
"Lang ",
"",
"",
"fu ",
"Bie ",
"",
"Tang ",
"",
"xiang ",
"",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ghet ",
"",
"Ngung ",
"",
"",
"",
"",
"",
"",
"",
"chang ",
"zhe ",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"kai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khan ",
"",
"",
"Ngay ",
"",
"",
"",
"Quo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"Khuay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hung ",
"",
"",
"",
"",
"",
"chu ",
],
"x229": [
"",
"sao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"zhan ",
"",
"",
"Cham ",
"",
"",
"",
"",
"",
"",
"zhuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Goe ",
"",
"",
"",
"",
"",
"",
"",
"",
"teng ",
"zhu ",
"",
"",
"Lung ",
"",
"Lo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trai ",
"",
"xie ",
"",
"jiao ",
"",
"",
"",
"Chong ",
"",
"Sung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Laan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"qiu ",
"Zai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x22a": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"zei ",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"hu ",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"Quanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ho ",
"",
"",
"",
"",
"",
"shan ",
"",
"Muon ",
"",
"Zit ",
"Hat ",
"",
"",
"",
"",
"",
"Thuon ",
"Dat ",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"xia ",
"",
"",
"",
"",
"Chop ",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vuc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngat ",
"Vat ",
"Nang ",
"",
"",
"",
"",
"Muc ",
"",
"",
"",
"",
"",
"",
"zhang ",
"",
"",
"",
"",
"Keo ",
"",
"That ",
"Vun ",
"",
"",
"",
"",
"",
"Tha ",
"",
"",
"Kam ",
"jie ",
"",
"",
"",
"",
"Wing ",
"Trot ",
"",
"So ",
"",
"",
"",
"Trum ",
"Rung ",
"Quay ",
"",
"",
"",
"",
"Bac ",
"",
"",
"Haku ",
"",
"Ngung ",
"",
"",
"Lat ",
"",
"nu ",
"",
"",
"zhe ",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"nie ",
"",
"",
"",
"",
],
"x22b": [
"",
"",
"",
"",
"Bung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Muc ",
"",
"",
"",
"",
"",
"",
"Chui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tay ",
"Khuong ",
"Giang ",
"",
"",
"",
"",
"",
"",
"",
"",
"Vot ",
"",
"",
"Khep ",
"",
"",
"",
"",
"",
"nan ",
"",
"",
"",
"",
"dun ",
"",
"",
"Xoi ",
"",
"",
"",
"Dau ",
"",
"",
"Chou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thuoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xac ",
"",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quet ",
"",
"",
"",
"Giau ",
"Khuay ",
"",
"",
"Vom ",
"",
"Lan ",
"Dui ",
"Xoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"bang ",
"",
"Luk ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Co ",
"Ron ",
"",
"Chut ",
"Co ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quay ",
"",
"",
"",
"",
"",
"lu ,jue",
"",
"",
"",
"xie ",
"Dik ",
"",
"zhai ",
"",
"Ngaau ",
"",
"Co ",
"",
"",
"Va ",
"Quat ",
"Ngoi ",
"Khep ",
"Quay ",
"Huo ",
"",
"",
"Sap ",
"Buoc ",
"Ven ",
"",
"",
"Va ",
"Roc ",
"Sua ",
"",
"",
"",
"Lay ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"Sau ",
"wan ",
"xue ",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"mao ",
"",
],
"x22c": [
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Wo ",
"",
"",
"",
"",
"Gap ",
"Zung ",
"pan ",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"jia ",
"",
"",
"",
"jia ",
"",
"Boi ",
"",
"Gieo ",
"Waa ",
"",
"",
"",
"Dap6",
"",
"",
"",
"Cai ",
"Phung ",
"Xoi ",
"",
"Nhot ",
"",
"Sin ",
"",
"",
"Saak ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kek ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung ",
"",
"",
"",
"",
"",
"",
"",
"",
"Run ",
"",
"",
"",
"Laa ",
"",
"Rung ",
"Cau ",
"",
"",
"Gay ",
"",
"Cap ",
"Mai ",
"Mo ",
"",
"",
"",
"",
"",
"Cau ",
"Sang ",
"",
"",
"Cao ",
"",
"Sou ",
"Lou ",
"",
"",
"",
"die ",
"",
"",
"zhu ",
"",
"",
"Bat ",
"",
"",
"",
"Ngao ",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"lang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Saai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x22d": [
"",
"",
"",
"",
"",
"",
"",
"Chap ",
"Daat ",
"",
"Chop ",
"Chong ",
"",
"",
"",
"",
"Day ",
"",
"Phanh ",
"",
"",
"Ning ",
"",
"",
"Xay ",
"",
"",
"",
"Xau ",
"",
"Nhung ",
"",
"",
"",
"fen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ban ",
"",
"",
"",
"",
"lei ",
"xie ,jie",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cou ",
"",
"",
"yang ",
"",
"",
"",
"",
"Dui ",
"",
"",
"",
"",
"",
"",
"Paang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zao ",
"",
"Dan ",
"",
"Doi ",
"",
"",
"Don ",
"",
"",
"",
"",
"Khoi ",
"",
"Sum ",
"Quay ",
"",
"Don ",
"Cat ",
"Xap ",
"",
"",
"",
"Vot ",
"Ro ",
"",
"",
"",
"",
"",
"",
"",
"Sip ",
"",
"",
"",
"",
"",
"",
"",
"Saap ",
"",
"",
"",
"Niao ",
"guai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngung ",
"",
"",
"Cui ",
"Saau ",
"",
"",
"Die ",
"Loe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Maan ",
"",
"",
"",
"",
"",
"",
"bang ",
"",
"Bum ",
"",
"Dom ",
"Bung ",
"Ngoi ",
"",
"",
"Don ",
"",
"Nem ",
"Xan ",
"",
"",
"Tro ",
"Chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dui ",
"",
"",
"",
"Hang ",
"",
"",
"Vo ",
"liu ",
"",
"",
"",
"du ",
"",
"",
"",
"",
"jie ",
],
"x22e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lan ",
"",
"",
"",
"",
"",
"",
"Niao ",
"",
"cuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Choc ",
"Gai ",
"Mac ",
"Rung ",
"",
"Xe ",
"",
"",
"",
"",
"yu ",
"",
"Zaang6",
"",
"",
"",
"",
"",
"",
"",
"lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Keo ",
"Xau ",
"",
"",
"",
"Tum ",
"Suong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giam ",
"Que ",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"ta ",
"",
"Naan ",
"",
"",
"",
"",
"",
"",
"Bung ",
"",
"Bau ",
"",
"",
"",
"",
"",
"",
"",
"",
"Uon ",
"",
"chan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Man ",
"yu ",
"",
"",
"",
"",
"Chia ",
"Chia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"jie ,fu",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x22f": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"da ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"Duk ",
"",
"",
"",
"",
"",
"",
"",
"Va ",
"",
"",
"",
"",
"wei ",
"",
"",
"yu ",
"du ",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"kuai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wen ",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"man ",
"",
"",
"",
"",
"xie ",
"luan ",
"",
"",
"",
"cheng ",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x230": [
"lei ",
"",
"",
"",
"qun ",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"cheng ",
"",
"",
"Chong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Va ",
"",
"",
"",
"fu ",
"",
"",
"",
"san ",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"Lon ",
"",
"gan ",
"tan ",
"",
"",
"",
"",
"man ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"dang ",
"",
"zhuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
],
"x231": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Huang ",
"",
"",
"",
"yan ",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liang ",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khuya ",
"Khuya ",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"Tam ",
"Luc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Phoi ",
"Trua ",
"",
"xu ",
"",
"",
"xian ",
"",
"",
"",
"gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"",
"",
"",
"",
"Chang ",
"",
"Giay ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sao ",
"",
"",
"Bie ",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x232": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"die ",
"",
"",
"",
"",
"",
"Phoi ",
"",
"Mai ",
"",
"Ngay ",
"",
"",
"",
"Quat ",
"Ngay ",
"",
"Hong ",
"",
"bian ",
"",
"Tia ",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ui ",
"Trua ",
"",
"",
"",
"",
"",
"",
"sang ",
"",
"ang ",
"nai ",
"",
"",
"",
"",
"",
"Gou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"Se ",
"",
"Mo ",
"Chop ",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"sang ",
"xu ",
"kan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bay ",
"",
"Sao ",
"Toi ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"dan ",
"",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Som ",
"",
"",
],
"x233": [
"",
"Trua ",
"Trua ",
"",
"",
"",
"Trua ",
"",
"",
"Khuya ",
"",
"Som ",
"Rua ",
"",
"",
"de ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nang ",
"",
"",
"chun ",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xeo ",
"",
"",
"",
"",
"",
"Nguc ",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"Cui ",
"",
"Oam ",
"",
"cha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Luo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nian ",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"Trang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"pai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jan ",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"Zai ",
"Then ",
],
"x234": [
"",
"",
"",
"",
"",
"Nhum ",
"Thot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xop ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Queo ",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"Cung ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Roi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gian ",
"",
"",
"Sim ",
"",
"",
"",
"Nen ",
"",
"",
"",
"fu ",
"",
"",
"dian ",
"",
"",
"",
"qiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xoai ",
"Sao ",
"Cong ",
"",
"",
"",
"han ",
"kuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sha ,jie",
"",
"",
"Gou ",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lau ",
],
"x235": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cui ",
"Bap ",
"",
"Pha ",
"Xoi ",
"Ngoc ",
"",
"Chanh ",
"Nhai ",
"",
"Kang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"Gon ",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"pao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Noc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Peng ",
"Chay ",
"Chay ",
"",
"",
"Ca ",
"",
"",
"",
"",
"",
"Suot ",
"Trac ",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"jue ",
"",
"ang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"reng ",
"",
"",
"Chua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"yang ",
"",
"",
"",
"",
"Son ",
"",
],
"x236": [
"",
"Ca ",
"",
"Cay ",
"Thot ",
"",
"",
"",
"Son ",
"Cum ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nik ",
"",
"bang ",
"Seot ",
"",
"",
"",
"",
"",
"",
"",
"",
"chun ",
"",
"",
"Yi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nau ",
"Vai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rac ",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"Sop ",
"",
"",
"Chay ",
"Rui ",
"",
"",
"",
"",
"jie ",
"zhe ",
"hu ",
"",
"Sot ",
"Con ",
"Mam ",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"pao ",
"",
"",
"",
"ye ",
"",
"",
"lei ",
"",
"ru ",
"",
"",
"juan ",
"",
"Jaap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuan ",
"",
"",
"",
"jiang ",
"hao ",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xanh ",
"",
"Gu ",
"",
"Khay ",
"",
"Be ",
"",
"",
"",
"Then ",
"Tu ",
"hu ",
"",
"",
"",
"",
"",
"Dom ",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"die ",
"",
"zha ",
"",
"",
],
"x237": [
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"",
"ai ",
"",
"",
"",
"Lim ",
"Son ",
"",
"",
"",
"",
"",
"",
"",
"",
"wen ",
"",
"",
"",
"Chua ",
"",
"",
"",
"hun ",
"",
"",
"ai ",
"",
"",
"",
"Duoi ",
"",
"ta ",
"",
"",
"",
"gao ",
"",
"yu ,yu",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"10389.190,bian ",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khu ",
"Cuoi ",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"",
"Tram ",
"",
"",
"meng ",
"",
"lu ",
"tan ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mong ",
"",
"",
"",
"xiao ",
"",
"huai ",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dui ",
"",
"",
"Ran ",
"",
"",
"",
"yu ",
"Kyo ",
"",
"",
"",
"",
"",
"",
"mei ",
"",
"lian ",
"",
"",
"lao ",
"",
"",
],
"x238": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tham ",
"Pheo ",
"",
"Chua ",
"Chua ",
"Que ",
"Gau ",
"",
"liu ",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tram ",
"",
"",
"",
"",
"",
"",
"",
"Tram ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"cen ",
"",
"",
"",
"",
"Tram ",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"Vui ",
"",
"die ",
"",
"",
"",
"",
"Nen ",
"",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"qian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"an ",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"xie ",
"",
"da ",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"",
"",
],
"x239": [
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"wan ",
"",
"",
"",
"",
"",
"",
"wang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"guan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"ta ",
"",
"chu ",
"",
"",
"zhu ",
"",
"da ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chu ",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Laai ",
"du ",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"",
"dian ",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nao ",
"",
"",
],
"x23a": [
"luan ",
"",
"die ",
"",
"qia ",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"tao ",
"",
"",
"",
"",
"",
"",
"zu ",
"ma ",
"",
"",
"",
"",
"",
"",
"jiang ",
"xu ",
"",
"",
"Giuoc ",
"",
"",
"",
"",
"",
"Quan ",
"",
"",
"",
"",
"du ",
"xiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"guan ",
"",
"",
"",
"",
"",
"er ",
"",
"",
"",
"",
"",
"",
"",
"liao ",
"",
"",
"",
"Ngoeo ",
"shan ",
"",
"",
"",
"",
"",
"zhai ",
"",
"ye ",
"diao ",
"",
"",
"",
"jiang ",
"",
"",
"",
"Toi ",
"huai ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ji ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"Van ",
"",
"dai ",
"",
"",
"",
"",
"",
"",
],
"x23b": [
"",
"",
"guan ",
"",
"",
"",
"pei ",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"juan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ba ",
"ba ",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"bao ",
"",
"",
"Su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"ge ",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ria ",
"Mao ",
"",
"",
"qiu ,qu",
"",
"",
"",
"",
"",
"",
"",
"Mau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngu ",
"",
"",
"de ",
"",
"jie ",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"san ",
"",
"chun ",
"",
"",
"",
"nai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"de ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"men ",
"",
"",
"",
"",
"",
"",
"",
"pei ",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"su ",
"",
"",
"",
],
"x23c": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"kun ",
"",
"",
"",
"",
"",
"jie ",
"",
"qu ",
"qu ",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"kun ",
"",
"",
"",
"Ho ",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"xiao ,hao",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hoi ",
"",
"",
"",
"Ram ",
"",
"",
"liu ",
"",
"",
"Vuc ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhop ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngut ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"Nap6",
"",
"",
"",
"",
"",
"",
"guai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"jie ",
"",
"",
"",
"",
"zhan ",
"",
"",
"",
"Deoi ",
"",
"",
"",
"",
"piao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Doe ",
"",
"",
"",
],
"x23d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"Hup ",
"Nhung ",
"",
"",
"",
"",
"",
"",
"zao ",
"zhuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"tan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung ",
"",
"",
"",
"",
"",
"Ngam ",
"",
"",
"",
"zhu ",
"",
"",
"",
"gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuang ",
"",
"",
"pao ",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Loi ",
"",
"Nhom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x23e": [
"",
"",
"",
"xu ",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"huai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"she ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lut ",
"",
"Tran ",
"",
"",
"",
"",
"",
"qiao ,xiao",
"",
"Cong ",
"",
"",
"",
"",
"",
"qian ",
"",
"",
"",
"xie ",
"",
"",
"hu ",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tao ",
"",
"qiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nuot ",
"",
"",
"Bui ",
"",
"",
"",
"Xoi ",
"",
"Duoi ",
"",
"",
"",
"dang ,xiang",
"",
"",
"",
"",
"",
"",
"",
"Ma ",
"",
"",
"",
"",
"shu ",
"",
"fu ",
"",
"",
"",
"xie ",
"lang ",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x23f": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"ze ",
"shuai ",
"",
"",
"Bot ",
"",
"",
"",
"",
"Vui ",
"Lung ",
"Ngau ",
"Doi ",
"Xop ",
"",
"",
"",
"Lot ",
"",
"",
"",
"",
"Tran ",
"Lang ",
"",
"",
"Ngau ",
"",
"",
"",
"",
"",
"Veo ",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"an ",
"jian ",
"",
"53066.030,teng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Doi ",
"",
"",
"",
"Khoi ",
"",
"",
"Xoi ",
"Bui ",
"",
"",
"",
"",
"Ngau ",
"",
"",
"fu ",
"",
"su ",
"",
"lian ",
"",
"he ",
"",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x240": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Loc ",
"",
"",
"Ngoi ",
"Rua ",
"Vung ",
"",
"",
"",
"Tanh ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dao ",
"Khoi ",
"",
"",
"",
"Trong ",
"Bot ",
"",
"Chua ",
"",
"",
"",
"",
"Dao ",
"dan ",
"jie ",
"bai ",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuan ",
"",
"",
"",
"",
"Rua ",
"Dan ",
"Phun ",
"Loi ",
"Toe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lan ",
"",
"yao ",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Leo ",
"",
"Muong ",
"",
"",
"",
"",
"",
"",
"",
"",
"Thuot ",
"lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Deng ",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
],
"x241": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"pan ",
"",
"",
"",
"",
"",
"",
"",
"Trong ",
"Nhan ",
"",
"Can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"man ",
"",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Veo ",
"",
"",
"",
"",
"",
"",
"men ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gua ",
"",
"xu ",
"",
"",
"",
"",
"Saap6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kho ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Say ",
"Phoi ",
"Tat ",
"",
"",
"",
"",
"Bep ",
"",
"",
"",
"Nhum ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ying ",
"",
],
"x242": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lui ",
"",
"Sot ",
"Ngut ",
"",
"",
"",
"",
"",
"he ",
"",
"",
"Cho ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lao ",
"shao ",
"",
"",
"",
"",
"Tro ",
"",
"Tro ",
"Se ",
"Heo ",
"Ngun ",
"",
"",
"Toa ",
"Rang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"kai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Luoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Roi ",
"",
"",
"Tro ",
"Nhem ",
"",
"",
"",
"",
"Rom ",
"",
"Phoi ",
"Phoi ",
"Lom ",
"",
"",
"Ben ",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"Ranh ",
"Nau ",
"Khet ",
"Kho ",
"",
"Phoi ",
"Kho ",
"Choi ",
"Um ",
"",
],
"x243": [
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hok ",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hay ",
"",
"",
"Ngot ",
"Nau ",
"",
"Ngun ",
"",
"",
"",
"",
"",
"",
"",
"Chong ",
"",
"shu ",
"",
"Jau ",
"",
"",
"",
"",
"",
"",
"zao ",
"",
"",
"",
"Jit6",
"",
"",
"",
"Zhang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Loa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quac ",
"",
"",
"",
"",
"Phap ",
"",
"",
"Tat ",
"",
"",
"",
"",
"",
"Ram ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhoi ",
"",
"Ho ",
"Thui ",
"Khet ",
"Thap ",
"Se ",
"",
"Rao ",
"Buot ",
"",
"",
"",
"",
"",
"chu ",
"Zhou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x244": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ham ",
"Nau ",
"",
"Soi ",
"",
"",
"Luoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"kai ",
"",
"",
"",
"",
"",
"",
"",
"",
"cuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"Chang ",
"",
"",
"",
"",
"Hung ",
"",
"",
"",
"",
"",
"",
"",
"Xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung ",
"Loe ",
"",
"",
"Sem ",
"",
"",
"",
"",
"",
"",
"",
"ye ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nung ",
"Nau ",
"",
"yao ",
"",
"Nhui ",
"",
"",
"Lom ",
"",
"",
"",
"Sem ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thap ",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ram ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ben ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vau ",
"",
"",
"",
"dao ",
"",
],
"x245": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shang ",
"Su ",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vuot ",
"",
"",
"",
"",
"",
"",
"",
"Danh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jiang ",
"",
"Jiang ",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"ju ",
"",
"",
"",
"Lop ",
"",
"",
"",
"xiang ",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"Mui ",
"",
"Sip ",
"",
"",
"",
"",
"",
"",
"pei ",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"Tam ",
"",
"",
"pu ",
"",
"",
"",
"Phuon ",
"",
"chan ",
"",
"",
"",
"",
],
"x246": [
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nga ",
"",
"",
"",
"",
"",
"",
"",
"Nanh ",
"hu ",
"",
"",
"",
"",
"",
"",
"chun ",
"",
"",
"",
"",
"tian ",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuang ",
"",
"",
"hu ",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"xie ",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngau ",
"",
"",
"",
"",
"",
"Caau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"Nghe ",
"",
"",
"die ",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"mei ",
"",
"Co ",
"Sao ",
"",
"",
"",
"",
"",
"",
"",
"xiu ",
"",
"",
"",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"Nghe ",
"",
"Trau ",
"",
"ceng ",
"ta ",
"",
"",
"jue ",
"xun ",
"",
"",
"",
"",
"",
"qun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x247": [
"",
"",
"",
"",
"huai ",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"ba ",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"San ",
"Tu ",
"",
"Cop ",
"",
"",
"",
"",
"",
"21335.010,yan ",
"",
"hu ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"Muop ",
"na ",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"Chuot ",
"",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"San ",
"Thac ",
"Tay ",
"",
"",
"",
"",
"",
"zang ",
"",
"",
"yu ",
"",
"cha ",
"xie ",
"",
"",
"",
"",
"Moi ",
"Khon ",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"hu ",
"zai ",
"jue ",
"",
"han ",
"",
"",
"",
"",
"Hum ",
"",
"",
"",
"an ",
"zao ",
"",
"",
"sha ",
"",
"xian ",
"",
"",
"",
"an ",
"",
"",
"",
"zhe ",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"xia ",
"xiao ",
"",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x248": [
"",
"",
"",
"",
"Ga ",
"Voi ",
"",
"ge ",
"",
"",
"",
"Trau ",
"",
"",
"",
"",
"ta ",
"Lau ",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"Gau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nanh ",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ran ",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"su ",
"",
"",
"",
"",
"Vuot ",
"San ",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ,qu",
"Nhen ",
"Dou ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"xie ",
"",
"",
"Nhat ",
"",
"",
"",
"",
"",
"",
"Beo ",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Doc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x249": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"da ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wai ",
"",
"",
"",
"",
"Vua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xa ",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"Cong ",
"",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x24a": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"xiang ",
"",
"",
"hao ",
"",
"",
"",
"dian ,tian",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"Dua ",
],
"x24b": [
"Dua ",
"",
"",
"",
"",
"",
"",
"",
"Hau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lu ",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"wa ",
"zu ",
"fan ",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"na ",
"Sanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"Ang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lei ",
"heng ",
"",
"Be ",
"",
"",
"",
"",
"",
"Lo ",
"",
"",
"lei ",
"",
"shan ",
"",
"",
"",
"Muong ",
"",
"",
"",
"",
"",
"",
"lu ",
"Lung ",
"",
"jun ",
"",
"",
"Chan ",
"",
"xie ",
"",
"zhe ",
"",
"",
"",
"",
"",
"liu ",
"lei ",
"",
"",
"",
"dai ",
"",
"Ngot ",
"Ngot ",
"",
"",
"",
"",
"",
"",
"Ngon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"Song ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x24c": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hang ",
"",
"",
"",
"",
"",
"",
"zha ",
"",
"",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"na ",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"Duoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ruong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cha ",
"Trai ",
"Trai ",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trong ",
"",
"",
"",
"",
"",
"Roc ",
"",
"",
"",
"",
"",
"",
"zhu ",
"Ve ",
"Ruong ",
"",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ruong ",
"",
"",
],
"x24d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngat ",
"",
"",
"",
"",
"",
"",
"",
"Bot ",
"",
"huan ",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"wu ",
"",
"wen ",
"",
"",
"",
"",
"",
"",
"",
"To ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Si ",
"",
"",
"qia ",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tay ",
"",
"",
"",
"",
"",
"Chau ",
"",
"",
"Nhan ",
"Ben ",
"",
"",
"tun ",
"fu ",
"",
"",
"",
"",
"",
"zhuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"chuan ",
"",
"",
"",
"",
"",
"",
"Naa ",
"Guoc ",
"",
"Nghen ",
"Mang ",
"Mut ",
"",
"",
"",
"",
"fei ",
"jian ",
"",
"",
"",
"Wak6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"lun ",
"",
"",
"",
"dao ",
"de ",
"",
"",
"",
"",
"",
"la ",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"Mang ",
"Cek ",
"",
"",
"",
"Loet ",
"",
"",
"",
"Nhom ",
"",
"",
"Buou ",
"",
"",
"wai ",
"",
"",
"",
"",
"",
"",
],
"x24e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"fu ",
"Ngung ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hoen ",
"",
"",
"",
"yun ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"bu ",
"",
"qun ",
"",
"",
"",
"Naa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"Lit ",
"",
"",
"",
"",
"",
"Hen ",
"",
"Nhoc ",
"Choc ",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"teng ",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Buou ",
"Not ",
"",
"Ngo ",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zang ",
"",
"",
"shan ",
"",
"",
"Buot ",
"Gua ",
"Mun ",
"Khom ",
"",
"",
"",
"Buou ",
"",
"",
"",
"",
"",
"",
"juan ",
"lu ",
"",
"ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cum ",
"",
"Hom ",
"",
"Toi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhoi ",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x24f": [
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chan ",
"",
"",
"",
"",
"guan ,huan",
"",
"",
"",
"",
"",
"",
"jue ",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"Dang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"er ",
"",
"huan ",
"",
"Nguoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"Phau ",
"",
"",
"Trang ",
"bai ",
"lu ",
"",
"",
"",
"",
"nian ",
"",
"",
"",
"",
"zhu ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tram ",
"Vang ",
"",
"",
"",
"",
"",
"",
"miao ",
"",
"",
"",
"Ho ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"",
"Nguoi ",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Saai ",
"",
"cun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Pi ",
"nan ",
"",
"",
"",
"",
"",
"",
"",
"wa ",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"cheng ",
"",
"",
"Da ",
"han ",
"xiao ",
"",
"Zaap ",
"",
"",
"",
"",
"",
"Trong ",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
],
"x250": [
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giay ",
"",
"",
"",
"",
"",
"",
"liu ",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xue ,qiao",
"",
"la ",
"",
"",
"",
"",
"",
"",
"Jim ",
"",
"",
"",
"la ",
"",
"du ",
"Mo ",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"jie ",
"mang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngaau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jim ",
"",
"ju ",
"",
"",
"",
"",
"",
"Trom ",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mam ",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"Mam ",
"Mam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vuc ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhap ",
"",
"",
"",
"fan ",
"",
"",
"",
"chuan ",
"yao ",
"",
"",
"",
"du ",
"",
],
"x251": [
"",
"meng ",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"Cik ",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gap6",
"",
"",
"Nham ",
"Ngom ",
"Nhon ",
"",
"mie ",
"",
"Xue ",
"xu ,yu",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhon ",
"",
"",
"",
"",
"Laap ",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"fan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngaau ",
"Hau ",
"er ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chau ",
"",
"",
"wei ",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"Zong ",
"",
"Giuong ",
"",
"",
"Nho ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"wang ",
"",
"juan ",
"",
"",
"",
"",
"",
"",
"xie ",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nao ",
"",
"",
"",
"wan ",
"jiu ",
"",
"Ngop ",
"Dau ",
"Ru ",
"Le ",
"",
"",
"Quau ",
"Mang ",
"Tro ",
"Bet ",
"",
"",
"",
"",
"",
"Nhon ",
"",
"",
"",
"han ,qia",
"",
"",
"xu ",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"Gwat6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"",
"",
"",
"Him ",
"",
"",
"",
"",
"Tro ",
"",
"",
"",
"xu ",
],
"x252": [
"",
"",
"",
"",
"chuang ",
"",
"mao ",
"",
"",
"huan ",
"sha ",
"",
"",
"",
"",
"kuang ",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"la ",
"",
"lu ",
"",
"",
"",
"Sou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"leng ",
"",
"",
"",
"Ngap ",
"Chot ",
"Nguoc ",
"Nhon ",
"Nom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"guan ",
"ju ",
"",
"nai ",
"",
"ge ",
"",
"",
"",
"",
"ma ",
"teng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Let ",
"",
"Soc ",
"",
"han ,qia",
"",
"",
"",
"",
"Ma ",
"lu ",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"Xam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"Caau ",
"",
"",
"",
"",
"",
"",
"chan ",
"",
"",
"",
"",
"Ghe ",
"Zong ",
"",
"",
"",
"Chop ",
"Quac ",
"Nhan ",
"",
"",
"",
"",
"Nguoi ",
"",
"Mu ",
"",
"",
"",
"",
"guan ",
"",
"zun ",
"",
"xie ",
"",
"",
"",
"",
"Toet ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Caang ",
"",
"",
"",
"",
"Sa ",
"",
"",
"",
"",
"",
"Tre ",
"",
"",
"Trom ",
"",
"",
"Saau ",
"",
"",
"",
"",
"",
"",
"Xi ",
"",
"",
"",
"jia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lem ",
"Coi ",
"Ngam ",
"Him ",
"Lam ",
"",
"Soi ",
"",
"",
"",
"",
"",
"",
"",
],
"x253": [
"Xet ",
"",
"",
"",
"",
"",
"",
"",
"Trom ",
"",
"pan ",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lai6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khoe ",
"Len ",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"meng ",
"wei ",
"42521.120,meng ",
"",
"",
"",
"",
"",
"",
"Dim ",
"Ngam ",
"yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lom ",
"",
"",
"",
"",
"Cham ",
"",
"lei ",
"",
"",
"",
"Nheo ",
"",
"bian ",
"Ngom ",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"",
"",
"",
"",
"zhai ",
"",
"",
"",
"",
"",
"",
"ze ",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kai ",
"",
"wu ",
"",
"",
"",
"",
"ze ",
"",
"",
"yu ",
"zan ",
"",
"",
"",
"xu ",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cuan ",
"cuan ",
"cuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"tiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"kua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"an ",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Van ",
],
"x254": [
"",
"pu ",
"",
"",
"",
"",
"Van ",
"Ngan ",
"So ",
"Ngan ",
"",
"",
"",
"Ngan ",
"",
"Coc ",
"",
"",
"Cut ",
"",
"Van ",
"ya ",
"",
"",
"Shi ",
"",
"",
"",
"",
"",
"Mong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"hu ",
"",
"",
"Chai ",
"",
"",
"",
"",
"die ",
"",
"",
"",
"guai ",
"",
"",
"",
"",
"",
"ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bu ",
"",
"",
"Sinh ",
"",
"",
"",
"",
"Nao ",
"",
"",
"",
"",
"du ",
"guai ",
"",
"Ran ",
"",
"",
"",
"Loi ",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"lu ",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"kang ",
"",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mai ",
"",
"",
"",
"",
"Sanh ",
"",
"",
"",
"du ",
"",
"",
"jie ",
"",
"xian ,kan",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"Mai ",
"",
"",
"Canh ",
"",
"",
"",
"",
"",
"",
"",
"Tuo ",
],
"x255": [
"Gwang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"Ji ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ham ",
"",
"",
"",
"",
"Mai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"Quanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zao ",
"Cuoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"Chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nen ",
"",
"",
"qu ",
"",
"ca ",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"lan ",
"",
"",
"",
"",
"",
],
"x256": [
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"ru ",
"",
"",
"Lay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"feng ",
"lu ",
"",
"",
"",
"",
"",
"",
"zhuan ",
"",
"zhe ",
"",
"",
"lu ",
"",
"",
"",
"",
"jue ",
"liao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x257": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ao ",
"",
"",
"",
"",
"",
"Yan ",
"",
"",
"",
"",
"",
"",
"zan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Yi ",
"",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"she ",
"",
"",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gua ",
"",
"jie ",
"",
"he ,xie",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"Li ",
"",
"",
"jie ",
"",
"ba ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"Cay ",
"du ,zha",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"Giong ",
"",
"zun ",
"",
"ru ",
"Duo ",
"jiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"heng ",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"ku ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lui ",
"",
"",
"Bap ",
"",
"",
"",
],
"x258": [
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trau ",
"Giong ",
"Rom ",
"Rom ",
"",
"",
"",
"",
"chang ",
"",
"",
"liu ",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ji ",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"Nanh ",
"Mam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zun ",
"",
"Cau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"kuai ",
"",
"ge ",
"xie ",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"ran ",
"",
"",
"",
"",
"",
"",
],
"x259": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"yang ",
"",
"",
"",
"liu ",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"qia ",
"dian ",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hang ",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"Nup ",
"",
"",
"",
"",
"zhe ",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"Tam ",
"",
"qu ",
"",
"",
"",
"",
],
"x25a": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"fu ",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hoam ",
"",
"Chui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ca ",
"",
"",
"",
"",
"",
"zhu ",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"Trong ",
"Trong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"pang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dun ",
"",
"",
"",
"lu ",
"Deon6",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"huang ",
"shi ",
"",
"",
],
"x25b": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khau ",
"Nia ",
"",
"",
"",
"",
"Giang ",
"Kao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mang ",
"Mau ",
"",
"qiu ",
"dun ",
"",
"",
"",
"",
"",
"",
"",
"mang ",
"",
"",
"miao ",
"yuan ",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fei ",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"Mang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Buong ",
"Hum ",
"Bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hang ",
"",
"ju ",
"nian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Manh ",
"Tre ",
"Mui ",
"",
"",
"",
"",
"Toi ",
"Trum ",
"",
"dang ",
"du ",
"",
"ye ",
"",
"",
"",
"",
"",
"",
"pai ",
"",
"",
"",
"",
"",
"",
"jian ,sha",
"",
"",
"",
"Trau ",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
],
"x25c": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"shan ",
"liu ",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"zhai ",
"",
"",
"",
"nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ro ",
"",
"",
"",
"Ghi ",
"Gianh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"dan ",
"",
"hu ",
"",
"",
"",
"",
"lu ",
"chuan ",
"wu ",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"shuang ",
"fu ",
"ju ",
"",
"",
"diao ",
"wang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"Khay ",
"",
"",
"",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"suan ",
"",
"an ",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
],
"x25d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thung ",
"Mui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"ge ",
"ju ",
"tun ,dian",
"",
"qia ",
"",
"",
"",
"jian ",
"",
"",
"",
"suan ",
"",
"",
"",
"",
"",
"qiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"Toi ",
"Hom ",
"Ray ",
"",
"Lak6",
"Nong ",
"",
"",
"jie ",
"zhu ",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mung ",
"Thung ",
"",
"Gay ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Toi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"ban ",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"May ",
"cu ",
"",
"",
"",
],
"x25e": [
"",
"Nan ",
"",
"",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trum ",
"",
"",
"",
"",
"",
"shan ",
"yu ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tam ",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"Nap6",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lau ",
"",
"",
"",
"bu ",
"chu ",
"",
"liu ",
"Bot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"Tam ",
"",
"Lo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Com ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nam ",
"",
"",
"",
"bu ",
"No6",
"",
"May ",
"",
"",
"",
"",
"sa ",
"",
"",
"mian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"Bun ",
"",
"",
"",
"",
"",
"",
"",
],
"x25f": [
"zu ",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lep ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"Thung ",
"",
"",
"",
"",
"",
"",
"",
"",
"Com ",
"",
"xian ",
"jian ",
"",
"",
"",
"Men ",
"",
"",
"",
"Miao ",
"",
"",
"",
"",
"",
"",
"",
"",
"Phan ",
"",
"",
"Xia ",
"",
"",
"",
"",
"",
"",
"",
"niang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Men ",
"",
"zhu ",
"",
"",
"",
"",
"",
"reng ",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"hu ",
"",
"",
"",
"",
"",
"jue ",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Buoc ",
"",
"",
"Vuong ",
"Tim ",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giay ",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"Khau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pai ",
"",
"",
"",
"",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x260": [
"",
"",
"",
"",
"",
"",
"",
"Jing ",
"",
"",
"Dai ",
"",
"",
"",
"",
"Thua ",
"",
"",
"",
"",
"",
"",
"Kep ",
"",
"",
"",
"",
"",
"zhuang ",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"No ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"Thun ",
"Xe ",
"Cui ",
"Noi ",
"Keo ",
"Rang ",
"",
"Boi ",
"Nuoc ",
"",
"",
"",
"",
"zhuan ,juan,shuan",
"nian ",
"",
"",
"hua ",
"",
"",
"",
"",
"yan ",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"Go ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ro ",
"",
"",
"May ",
"Vuong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wan ",
"",
"",
"Sok ",
"",
"",
"",
"The ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Luot ",
"",
"",
"",
"Buoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tao ",
"",
"",
"liu ",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x261": [
"",
"Thun ",
"Dai ",
"",
"Quan ",
"",
"May ",
"",
"",
"",
"",
"",
"Chap ",
"",
"",
"",
"",
"",
"",
"Xau ",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Van ",
"",
"Nut ",
"",
"",
"Cuon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chai ",
"",
"",
"",
"Xung ",
"",
"Rang ",
"Quan ",
"",
"qu ",
"",
"",
"xun ",
"",
"",
"",
"zhe ",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tan ",
"",
"",
"Thua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Neo ",
"",
"",
"",
"wu ",
"Mung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Roi ",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yao ",
"",
"Chung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
],
"x262": [
"",
"",
"Nhau ",
"nang ",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cha ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Caang ",
"bian ",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"cen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"Bon ",
"Tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"fu ",
"liu ",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"han ",
"",
"",
"Ra ",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x263": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"su ",
"",
"",
"",
"",
"",
"",
"",
"liao ",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"nan ",
"",
"",
"",
"",
"quan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"ta ",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gu ",
"fen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"chuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"Guong ",
"chuan ",
"",
"",
"",
"",
"Xinh ",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"Tanh ",
"fu ",
"",
"",
"gu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ren ",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bon ",
"",
"du ",
"",
"hu ",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"",
"",
"huai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ,yu",
"",
"",
"",
"",
],
"x264": [
"",
"",
"",
"",
"",
"",
"",
"Taap ",
"Fen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ,yu",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"zu ",
"",
"xu ",
"yan ",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"ta ",
"",
"",
"fu ",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zin ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tai ",
"",
"",
"",
"",
"Khu ",
"shu ",
"Bou ",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gia ",
"",
"Khu ",
"",
"Lu ",
"",
"wang ",
"",
"",
"nai ",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ma ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x265": [
"",
"",
"",
"",
"",
"tu ",
"",
"",
"ze ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cay ",
"",
"",
"",
"",
"",
"",
"pai ",
"",
"",
"",
"",
"",
"kuai ",
"",
"",
"",
"",
"",
"qu ",
"",
"zhe ",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nai ",
"",
"",
"",
"",
"",
"tian ",
"",
"",
"",
"",
"",
"",
"ye ",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"Zim ",
"xu ",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"duo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hua ",
"",
"",
"",
"Nghe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tai ",
"hu ",
"",
"",
"Dap ",
"",
"tian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"Lang ",
"ai ",
"zhe ",
"",
"lu ",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"Ghe ",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
],
"x266": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wai ,wa",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Paa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Moc ",
"Mao ",
"Hong ",
"Tim ",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"dai ",
"",
"",
"",
"",
"zhu ",
"Wan ",
"Gu ",
"",
"",
"ban ",
"",
"mai ",
"Ci ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ping ",
"",
"",
"Zaap ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hen ",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ruot ",
"",
"",
"ku ",
"na ",
"",
"",
"",
"xuan ",
"",
"",
"",
"he ",
"",
"Nam6",
"",
"Ham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"Bet ",
"Thon ",
"",
"",
"Nuc ",
"Mang ",
],
"x267": [
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pang ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngaa6",
"",
"Vu ",
"",
"",
"",
"",
"",
"Ron ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rang ",
"",
"",
"",
"Ngac ",
"",
"",
"Mun ",
"Mep ",
"",
"Phop ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"",
"",
"",
"Jim ",
"",
"",
"ban ",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tuoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nghen ",
"",
"",
"",
"",
"Phay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tao ",
"Sao ",
"",
"zhe ",
"",
"",
"",
"",
"",
"sai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bo ",
"Zin ",
"Gay ",
"Suoi ",
"Khu ",
"",
"",
"",
"tun ",
"",
"",
"",
"Nem ",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"cu ",
"",
"",
"",
"Xiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hun ",
"ju ",
"",
"",
"Nem ",
"",
"",
"",
"",
"",
"",
"",
"Khu ",
"",
],
"x268": [
"",
"cu ",
"",
"",
"",
"xun ",
"Sun ",
"ceng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"pao ",
"",
"Vai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"zhe ",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Phet ",
"",
"",
"",
"Roi ",
"Seo ",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"du ",
"",
"Fat ",
"",
"",
"",
"",
"Ron ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngam ",
"Lung ",
"Song ",
"",
"Rau ",
"",
"",
"",
"",
"",
"san ",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"yao ,shao",
"",
"",
"",
"hun ",
"Lom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Pok ",
"",
"Duk6",
"",
"guang ",
"",
"",
"",
"",
"Mak6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rang ",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Vai ",
"Lot ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lei ",
"",
"",
"Lo ",
"",
"",
"",
"",
"",
"",
"Nan ",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"Nhau ",
"",
"Nang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zak ",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
],
"x269": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"yao ",
"",
"",
"Gao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hot ",
"zhai ",
"",
"",
"",
"",
"hai ",
"",
"Thoi ",
"Kham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"",
"",
"",
"",
"Kyo ",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ,fan",
"",
"",
"",
"",
"",
"Lau ",
"",
"Weng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tian ",
"",
"hu ",
"",
"",
"ta ",
"",
"",
"ta ",
"ta ",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"huai ",
"",
"",
"",
"",
"ta ",
"Loe ",
"",
"",
"hua ",
"",
"",
"zhuan ",
"",
"Laai ",
"",
"",
"",
"",
"",
],
"x26a": [
"",
"",
"Mua ",
"",
"",
"",
"",
"",
"fu ",
"wu ",
"",
"fu ",
"",
"",
"ta ",
"",
"",
"Chai ",
"",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"Qua ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"jia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"Khoang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mui ",
"Gu ",
"",
"",
"yu ",
"",
"",
"fu ",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sheng ",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"Ao ",
"",
"",
"",
"",
"",
"",
"Thong ",
"",
"",
"",
"",
"dun ",
"",
"",
"",
"jue ",
"ta ",
"zun ",
"",
"",
"teng ",
"",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ghe ",
"",
"",
"",
"",
"",
"Khoang ",
"Ghe ",
"",
"",
"",
"teng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ba ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wa ",
"",
"",
"",
"",
"xun ",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x26b": [
"",
"",
"wu ",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tup ",
"",
"",
"Hung ",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nen ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"Dua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x26c": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"Ngon ",
"",
"",
"Nua ",
"",
"Lau ",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
"dun ",
"",
"",
"",
"",
"",
"ye ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"yu ",
"",
"yu ",
"yu ",
"Gu ",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"Bei ",
"",
"Fu ",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"Lei6",
"",
"",
"",
"Rom ",
"",
"",
"Muop ",
"",
"",
"hao ,mao",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nai ,na",
"",
"",
"fu ",
"du ",
"",
"",
"",
"",
"Tre ",
"",
"",
"",
"",
"bai ",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"juan ",
"ru ",
"",
],
"x26d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"Mong ",
"jun ",
"she ",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ke ",
"",
"",
"Nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"heng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"qiang ",
"",
"",
"Bom ",
"",
"",
"",
"",
"",
"Mai ",
"",
"",
"",
"hu ",
"",
"",
"",
"hai ",
"ru ",
"meng ",
"",
"",
"",
"wu ",
"",
"",
],
"x26e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pei ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"Thom ",
"",
"",
"Voi ",
"Bui ",
"",
"Nhai ",
"",
"Dam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"chuan ",
"Lao ",
"",
"",
"hu ",
"",
"",
"jie ",
"xiang ",
"",
"",
"",
"",
"xiang ",
"",
"",
"lang ",
"",
"",
"",
"",
"",
"",
"shuan ",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zaau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x26f": [
"",
"",
"",
"Sang ",
"",
"",
"",
"",
"ju ",
"",
"leng ",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lum ",
"San ",
"San ",
"",
"Dua ",
"Khay ",
"Sung ",
"",
"she ",
"",
"",
"",
"",
"sa ",
"",
"",
"mao ",
"qu ",
"",
"",
"",
"Zuk6",
"",
"juan ",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"lu ",
"mian ",
"dian ",
"",
"",
"",
"",
"",
"Waa ",
"",
"",
"",
"lu ",
"fu ",
"",
"",
"zei ",
"",
"Om ",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tranh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"lu ",
"",
"Dua ",
"",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"lu ",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"chu ",
"fu ",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x270": [
"",
"",
"",
"",
"chang ",
"",
"meng ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"Sung ",
"",
"Giong ",
"Non ",
"Ru ",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"Sam ",
"",
"",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"Lop ",
"",
"",
"",
"",
"",
"",
"",
"",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"fan ",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ghem ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"jian ",
"han ",
"dan ",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"meng ",
"ju ",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dua ",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Muong ",
"Khoai ",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x271": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Han ",
"",
"",
"",
"",
"fu ",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Men ",
"Muong ",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thuoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"zu ",
],
"x272": [
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"42573.210,zhu ",
"",
"teng ",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tai ",
"",
"",
"",
"",
"",
"",
"",
"Trun ",
"Vat ",
"",
"",
"",
"",
"",
"",
"",
"xue ",
"yu ",
"fan ",
"",
"",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gwaai ",
"",
"dai ,de",
"",
"Buom ",
"Nhong ",
"",
"",
"",
"Vat ",
"",
"Trut ",
"",
"",
"",
"",
"",
"",
"",
"ku ",
"",
"",
"qu ",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"Gong ",
"Doe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Saa ",
"",
"du ",
"Zhe ",
"",
"",
"",
"",
"",
"kao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Moi ",
"",
"na ",
"",
"",
"",
"",
"Mei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cuong ",
"Mot ",
"",
"",
"",
"Chay ",
"",
"Nhong ",
"Mang ",
"",
"",
],
"x273": [
"",
"",
"Bo ",
"",
"",
"",
"",
"Bang ",
"fu ",
"lu ",
"",
"",
"",
"lu ",
"",
"ta ",
"",
"",
"",
"Fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cua ",
"",
"",
"",
"",
"Ngai ",
"",
"",
"",
"Cua ",
"",
"",
"",
"shuai ",
"",
"jue ",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"yu ",
"",
"",
"",
"",
"feng ",
"",
"die ",
"",
"",
"lian ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chau ",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"Nhuc ",
"Sam ",
"",
"",
"Bo ",
"Choi ",
"",
"",
"Nhuc ",
"",
"ai ",
"",
"",
"",
"",
"",
"zhu ",
],
"x274": [
"",
"su ",
"",
"xie ",
"yu ,yu",
"Zeoi ",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"Luon ",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Hou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"lu ",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bo ",
"Sung ",
"",
"Sam ",
"",
"Ngao ",
"",
"Vet ",
"Chang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"Giun ",
"",
"dai ",
"",
"",
"",
"dang ",
"zu ",
"",
"",
"",
"chuan ",
"",
"",
"du ",
"",
"",
"",
"",
"",
"xie ",
"zhe ",
"",
"",
"",
"sao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"Buom ",
"",
"Sau ",
"",
"Rua ",
"",
"Rom ",
"Rua ",
"Rua ",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lei ",
"",
"zei ",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"fei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngoe ",
"Sam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x275": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"qiu ",
"",
"",
"",
"",
"",
"",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trai ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"chai ",
"zang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Naan ",
"",
"",
"",
"",
"",
"shuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Moi ",
"",
"",
"",
"",
"",
"",
"",
"an ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"n ",
"",
"",
"san ",
"hu ",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Jiku ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"yu ,qu",
"jue ",
"",
"",
"",
"",
"",
"shu ,yu",
"",
"",
"Jung ",
"",
"",
"Jutsu ",
"jian ",
"",
"",
"",
"",
"",
"",
"shuai ",
],
"x276": [
"",
"",
"Chong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Taai ",
"",
"",
"",
"",
"",
"",
"La ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pu ",
"",
"che ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"53077.090,jian ",
"",
"",
"",
"",
"",
"zhan ",
"yuan ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"Lot ",
"",
"",
"",
"",
"mu ",
"huan ",
"",
"",
"e ",
"Long6",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"Tung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"",
"Xong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trang ",
"Lun ",
"May ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"tu ",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Toi ",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x277": [
"",
"Sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chan ",
"Mo ",
"",
"Xong ",
"",
"",
"",
"",
"",
"",
"shuai ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bau ",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"heng ",
"",
"qu ",
"wei ",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"Gei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cheo ",
"Toang ",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"meng ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"La ",
"shu ",
"jie ",
"lei ",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x278": [
"",
"",
"",
"",
"",
"",
"",
"",
"cha ",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"chan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thay ",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lao ",
"",
"",
"",
"",
"",
"Don ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lai6",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Caau ",
"",
"",
"",
"qiao ",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"Dia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lun ",
"",
"",
"jue ",
"",
"",
"ju ",
"hu ",
"",
"",
"",
],
"x279": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ba ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thoi ",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"Choi ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"kuang ",
"",
"jue ",
"",
"",
"",
"",
"",
"Va ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"fu ",
"",
"",
"yang ",
"",
"",
"ren ",
"",
"",
"Kwan ",
"",
"yun ",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pu ",
"",
"Ngaak ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Beng ",
"Ngoa ",
"",
"",
"",
"wang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngaa6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Khoang ",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"bie ",
"",
"",
"zha ",
"",
"",
"",
"",
],
"x27a": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zaa6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"Tam ",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Daan ",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"la ",
"",
"",
"",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ,ge",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ma ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"jie ",
"xia ",
"",
"",
"",
"",
"cha ",
"",
"",
"",
"yang ",
"",
"",
"",
"",
"",
"",
"wang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mo ",
"",
"jiu ",
"",
],
"x27b": [
"",
"hao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Wai ",
"",
"gun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"teng ",
"",
"",
"",
"",
"",
"",
"",
"Leoi6",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"he ",
"",
"",
"",
"",
"zhan ",
"",
"tai ",
"qian ",
"",
"",
"",
"",
"",
"",
"Hang ",
"Hang ",
"",
"",
"",
"",
"",
"",
"",
"lao ",
"",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zai ",
"juan ",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"kan ",
"",
"",
"",
"",
"Phong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x27c": [
"",
"",
"",
"",
"",
"",
"",
"yu ",
"wu ",
"",
"Tian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Shuu ",
"",
"",
"han ",
"chu ",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"na ",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"lun ",
"jue ,jun",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ceng ",
"",
"hu ",
"",
"",
"",
"ju ",
"sha ",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"Cop ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"Beo ",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"Cop ",
"",
"",
"",
"",
"",
"",
"wu ",
"pei ",
"",
"",
"",
"",
"",
"",
"",
"",
"Hum ",
"",
"",
"sha ",
"",
"zhao ",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tuan ",
"",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gu ",
"shao ",
"",
"",
"",
"",
"",
"",
],
"x27d": [
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"Beo ",
"fu ",
"",
"biao ",
"",
"",
"",
"",
"",
"",
"biao ",
"",
"",
"",
"",
"guai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pei ",
"",
"",
"",
"",
"Suo ",
"",
"",
"",
"",
"Me ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Qua ",
"pai ",
"",
"",
"",
"",
"",
"",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"She ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Danh ",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cha ",
"Faan ",
"ze ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fen ",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xoe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"Mua ",
"",
"",
"",
"",
"",
"ju ",
"",
],
"x27e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zhuan ",
"xue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lan ",
"ju ",
"",
"xun ",
"zhan ",
"gun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bui ",
"",
"",
"",
"chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"reng ",
"",
"",
"Vay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Son ",
"",
"",
"xu ",
"",
"",
"",
"Tham ",
"hu ",
"gan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"Tham ",
"Tham ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"ju ",
"",
"Mut6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ,zao",
"qiu ",
"",
"fu ",
"lang ",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x27f": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lanh ",
"",
"",
"",
"xu ",
"",
"",
"leng ",
"",
"",
"fu ",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"jie ,jue",
"",
"",
"yu ",
"",
"",
"Tang ",
"shu ,yu",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"Day ",
"Day ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"lu ",
"",
"chu ",
"",
"",
"",
"lian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"e ",
"su ",
"Jue ",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"Tron ",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"xun ",
"",
"",
"xun ",
"",
"",
"ju ",
"",
"du ",
"",
"",
"",
"xun ,xuan",
"",
"",
"",
"",
"",
"jie ",
"",
"qu ",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"",
"Treo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ku ,wu",
"",
"ku ",
"zha ",
"",
"",
"ba ",
"Chen ",
"Nhac ",
"Dam6",
"hu ",
"nu ",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"ge ",
"",
"Dau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"Mau ",
"Xoat ",
"ju ",
"",
"",
"",
"",
],
"x280": [
"",
"",
"",
"",
"",
"",
"",
"",
"Buoc ",
"",
"",
"",
"",
"Dung ",
"Lop ",
"",
"Rong ",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"cun ",
"",
"",
"",
"",
"",
"",
"",
"Kei ",
"Choi ",
"",
"",
"",
"",
"",
"",
"",
"zai ",
"",
"",
"",
"Bang ",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"xie ",
"",
"han ",
"",
"",
"tun ",
"Gaang ",
"",
"cen ",
"",
"",
"Ren ",
"Choanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Buot ",
"",
"",
"",
"Noi ",
"",
"",
"",
"",
"",
"Noi ",
"Cuong ",
"",
"Dau ",
"",
"",
"",
"",
"",
"Xui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bon ",
"de ",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Leoi ",
"",
"",
"",
"Doc ",
"",
"",
"Co ",
"",
"",
"",
"",
"",
"",
"tan ",
"",
"",
"",
"",
"wu ",
"",
"",
"chuan ",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dam6",
"Naam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giay ",
"Tot ",
"",
"Xam ",
"Giay ",
"",
"",
"",
"",
"Soc ",
"kua ",
"teng ",
"",
"",
"ta ",
"sa ",
"",
"",
"Pun ",
"Pun ",
"",
"",
"",
"sao ",
"",
"",
"",
"",
"",
"Sin ",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"jie ",
"neng ",
"",
"",
"",
"Chuc ",
"",
"",
"",
"",
"To ",
"Nhuc ",
"Xung ",
"",
],
"x281": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Te ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shuan ",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"Sup ",
"",
"",
"",
"",
"",
"Chui ",
"",
"",
"",
"Ghe ",
"Giong ",
"",
"Sup ",
"Mop ",
"",
"",
"",
"",
"",
"",
"Lung ",
"",
"xiao ",
"",
"Ren ",
"",
"Laam ",
"",
"",
"shu ,chu",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jaang ",
"Nhap ",
"",
"",
"",
"",
"",
"",
"Bay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Doi ",
"",
"",
"",
"",
"",
"",
"Sum ",
"",
"",
"",
"Khum ",
"",
"zhu ",
"",
"",
"cha ",
"juan ",
"",
"",
"",
"",
"",
"zei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cui ",
"",
"",
"Quay ",
"Dep ",
"",
"",
"",
"Tuon ",
"",
"",
"",
"",
"Jaang ",
"",
"",
"",
"",
"Buk6",
"ta ,da",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuang ",
"",
"",
"",
"bao ",
"lai ",
"",
"",
"Leo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chan ",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhao ",
"Khuy ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x282": [
"",
"",
"",
"",
"",
"",
"",
"Laam ",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"leng ",
"",
"",
"",
"",
"",
"",
"Pei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ye ",
"",
"",
"",
"",
"",
"",
"huang ",
"",
"",
"",
"",
"",
"",
"",
"Pei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nei ",
"",
"",
"",
"",
"",
"",
"",
"Lung ",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mei ",
"",
"",
"",
"tang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Wu ",
"xiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Laak ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"xuan ",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"Wan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zha ",
"",
"",
"pao ",
"",
"",
"bu ",
"he ",
"",
"",
"Lip ",
"",
"So ",
"",
"",
"ju ",
"hun ",
"",
"",
"",
"",
"",
"",
"Zhuai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x283": [
"",
"",
"",
"",
"zang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"Tang ",
"",
"",
"chao ",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Co ",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"xian ,jian",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sau ",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"Yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"So ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Truoc ",
"Truoc ",
"",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"jian ",
"jiao ",
"",
"",
"",
"",
"",
"",
"peng ",
"Dang ",
"",
"qu ",
"",
"mu ",
"",
"",
"",
"",
"",
"fen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shuan ",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ,du",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"So ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"ge ",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
],
"x284": [
"",
"die ",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ku ",
"",
"qu ",
"",
"ge ",
"ban ",
"",
"",
"",
"",
"",
"",
"",
"Cay ",
"",
"",
"ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ban ,bian",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"Tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xuong ",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pan ",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"kan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"Lung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pei ",
],
"x285": [
"",
"",
"",
"",
"",
"",
"",
"",
"Choi ",
"",
"Suot ",
"",
"Co ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chuc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"chu ",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"Sang ",
"",
"",
"",
"",
"",
"",
"yang ",
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"xiang ",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhanh ",
"yu ,ju",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"da ",
"",
],
"x286": [
"",
"",
"jiu ",
"",
"",
"",
"",
"",
"",
"sha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiu ",
"",
"",
"Nhanh ",
"",
"",
"kao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Keo ",
"ta ",
"zan ",
"",
"",
"",
"",
"zhu ",
"",
"Suot ",
"",
"",
"",
"",
"",
"",
"Lui ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qie ",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"",
"cun ",
"",
"ru ",
"zai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shao ",
"han ",
"",
"jun ",
"",
"bu ",
"",
"",
"",
"kuai ",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"pei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x287": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"huan ",
"qiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chuang ",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhai ",
"",
"",
"",
"",
"",
"guan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"shun ",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"dang ",
"",
"",
"",
"",
"",
"",
"feng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"yu ",
"",
"zhu ,chu",
"",
"qun ",
"",
"qu ",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gai ",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"wan ",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
],
"x288": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Tam ",
"",
"",
"",
"",
"",
"",
"",
"",
"Faan ",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"",
"",
"",
"Giam ",
"Gay ",
"Sua ",
"Xoang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"Dam ",
"",
"",
"nei ",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"zha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"Gay ",
"",
"",
"hun ",
"",
"mu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"han ",
"",
"meng ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cay ",
"meng ",
"",
"mu ",
"",
"hu ",
"Mi ",
"shai ,zha",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"nian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chuai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"meng ",
"pao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Giam ",
"Dau ",
"Zeoi6",
"",
"lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x289": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"juan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"guang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"cen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Keo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nen ",
"",
"",
"",
"",
"",
"Lao ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiang ",
"",
"",
"",
"bian ",
"",
"wu ",
"",
"",
"",
"",
"Bong ",
"",
"",
"",
"Gang ",
"",
"",
"Xot ",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x28a": [
"",
"",
"",
"",
"",
"Hoat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"quan ",
"",
"",
"chao ",
"he ",
"",
"",
"",
"",
"Cuoc ",
"wu ",
"",
"",
"Khep ",
"",
"Mai ",
"Thep ",
"",
"",
"",
"",
"",
"",
"",
"",
"ruan ",
"",
"",
"zu ",
"",
"",
"yu ",
"tu ",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mao ",
"",
"",
"yu ",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ruan ",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bam ",
"Lon ",
"",
"",
"Choc ",
"",
"Sat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cuoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x28b": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dui ",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"ruan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"Baang ",
"Bay ",
"",
"Choang ",
"",
"",
"Choang ",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Thau ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Com ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung ",
"Hom ",
"Cun ",
"",
"",
"",
"",
"",
"",
"Nhon ",
"Thoi ",
"",
"",
"yu ",
"",
"",
"",
"Ban ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Dui ",
"",
"",
"",
"",
"",
"",
"",
],
"x28c": [
"",
"",
"qian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"diao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nen ",
"",
"",
"",
"yu ",
"",
"Ben ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"duan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cen ",
"Dai ",
"Dai ",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"zhu ",
"",
"an ",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"feng ",
"wu ",
"jiao ",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"niao ",
"",
"chan ",
"",
"",
"",
"nang ",
"",
"",
"",
"Gau ",
"",
"",
"Cat6",
"",
"",
"",
"",
"Mon ",
"",
"men ",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
],
"x28d": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"Kuang ",
"",
"wei ",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"he ",
"yan ",
"",
"",
"Cua ",
"",
"",
"tu ",
"",
"",
"hu ",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"men ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"ta ",
"",
"fu ",
"",
"",
"",
"Cua ",
"",
"",
"yu ",
"",
"",
"",
"xie ",
"",
"xian ",
"jian ",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lang ",
"",
"Lan ",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hua ",
"wen ",
"",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gwaan ",
"",
"",
"",
"",
"",
"",
"",
"niu ",
"",
"",
"xiang ",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"run ",
"",
"",
"",
"",
"jian ",
"xu ",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cua ",
"",
"",
"",
"",
"",
"",
"Cua ",
"",
"",
"",
"",
"",
"",
"",
"dang ",
"",
],
"x28e": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cup ",
"",
"",
"reng ",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"Ji ",
"gai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zhen ",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"duan ",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"Luong ",
"",
"",
"",
"",
"",
"",
"Xia ",
"",
"",
"",
"Geki ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x28f": [
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"cuan ",
"wu ",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"Be ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"zhan ",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"meng ",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"fang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x290": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"run ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cai ",
"",
"",
"",
"",
"",
"",
"",
"Loi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Con ",
"Ngut ",
"",
"shai ",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"Che ",
"Ram ",
"Bung ",
"",
"",
"fu ",
"Phat ",
"ze ",
"pu ",
"",
"",
"",
"",
"",
"",
"pao ",
"Mu ",
"",
"",
"",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rei ",
"",
"dian ",
"Set ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ba ",
"",
"",
"dai ",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x291": [
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"Mua ",
"Bung ",
"",
"",
"",
"",
"",
"dan ",
"meng ",
"xia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"peng ",
"",
"Guot ",
"",
"",
"",
"",
"",
"",
"wu ",
"Set ",
"",
"May ",
"",
"",
"Nap ",
"Xoi ",
"piao ",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"man ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung ",
"",
"",
"",
"",
"nan ",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cen ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"May ",
"Rau ",
"Zyu ",
"",
"",
"dan ",
"fu ",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"Rao ",
"",
"Rao ",
"liu ",
"",
"Sam ",
"",
"",
"",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mong ",
"Rao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sam ",
"",
"",
"",
"",
"",
"",
"Khuya ",
"",
"",
"",
"",
"",
"",
"Mu ",
"Sam ",
"hun ",
"",
"",
"",
"",
"",
"Loa ",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xanh ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fei ",
"",
"",
],
"x292": [
"",
"",
"Bay ",
"",
"",
"",
"",
"",
"pang ",
"dan ",
"",
"ai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"wan ",
"",
"diao ",
"",
"",
"",
"suan ",
"",
"",
"",
"",
"",
"",
"mian ",
"",
"",
"",
"",
"",
"",
"lan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gan ",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"hang ",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ang ",
"",
"",
"",
"",
"fen ",
"",
"",
"ju ",
"",
"",
"",
"",
"fu ",
"",
"qu ",
"",
"",
"",
"ma ",
"",
"bao ",
"",
"yu ",
"",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"bu ",
"",
"",
"",
"jue ",
"xun ",
"",
"Hia ",
"",
"",
"",
"",
"bai ",
"",
"",
"ta ",
"",
"",
"nao ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xie ",
"diao ",
"",
"",
"",
"",
"die ",
"fu ,fu",
"",
"",
"xuan ,yun",
"",
"yu ",
"",
"",
"Xie ",
"fu ",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"gao ",
"",
"",
"e ",
"",
"mei ",
],
"x293": [
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ",
"lu ",
"",
"",
"xu ",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"",
"",
"",
"sheng ",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"Bang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Roi ",
"Dep ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"",
"",
"yu ",
"",
"bai ",
"",
"ta ",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"yun ",
"duan ",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hun ",
"",
"",
"",
"",
"",
"",
"",
"bai ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"juan ",
"jue ",
"",
"",
"",
"",
"",
],
"x294": [
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"peng ",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"e ",
"",
"",
"",
"",
"",
"",
"guang ",
"",
"",
"",
"",
"ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"men ",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"yao ",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"wu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"Ngup ",
"Mang ",
"bie ",
"",
"",
"",
"",
"",
"",
"an ",
"Ngok6",
"wu ",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
],
"x295": [
"wai ",
"",
"dun ",
"",
"jie ",
"",
"",
"",
"",
"",
"zhuan ",
"hang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qiu ",
"",
"Lei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liao ",
"",
"",
"",
"",
"wai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"dan ",
"",
"",
"",
"",
"",
"",
"",
"",
"Gai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"To ",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"",
"hua ",
"",
"",
"",
"",
"qiu ",
"",
"",
"",
"",
"",
"",
"",
"",
"sa ",
"",
"",
"",
"",
"",
"han ",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bung6",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"",
],
"x296": [
"zhao ",
"",
"",
"",
"May ",
"",
"",
"",
"",
"",
"la ",
"",
"",
"yuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"To ",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"shuai ",
"",
"yu ",
"",
"",
"",
"",
"",
"su ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"niu ",
"",
"",
"",
"Bay ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nhu ",
"",
"niu ",
"",
"",
"",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Qua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"No ",
"Qua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zai ",
"",
"",
"mao ",
"",
"yu ",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"lu ",
"",
"",
"",
],
"x297": [
"",
"",
"",
"ju ",
"",
"",
"",
"Juan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ang ",
"Mam ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wu ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Qiu ",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chan ",
"",
"jiang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ",
"",
"yuan ,xuan",
"Qua ",
"",
"",
"",
"",
"ao ",
"",
"",
"",
"shan ",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"Caat ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"huai ",
"",
"",
"",
"",
"yu ",
"",
"",
"chan ,jie",
"",
"",
"",
"",
"Nang ",
"",
"",
"",
"",
],
"x298": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ding ",
"",
"",
"Mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cui ",
"",
"",
"",
"",
"",
"",
"Guc ",
"",
"",
"",
"",
"",
"",
"",
"wei ",
"",
"Chui ",
"",
"Chui ",
"Cui ",
"Choi ",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"Thom ",
"",
"",
"",
"",
"",
"",
"",
"",
"hai ",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wan ",
"",
"",
"",
"",
"",
"",
"",
"wen ",
"",
"",
"",
"zhe ",
"",
"",
"",
"ban ",
"bu ",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"Giong ",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"fu ",
"",
"Sai ",
"qu ",
"",
"",
"yu ",
"",
"",
"jiu ",
"",
"shu ",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"Ngon ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"",
"tu ",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
],
"x299": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ,jian",
"kun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"tan ",
"",
"xie ",
"",
"",
"duan ",
"",
"",
"",
"",
"",
"e ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"",
"",
"",
"Au6",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"tu ",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bang ",
"yu ",
"Jyu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xian ",
"meng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cai ",
"du ",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x29a": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"yu ",
"",
"",
"ba ",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"fan ",
"",
"",
"",
"Que ",
"",
"",
"",
"",
"ma ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kheo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"",
"",
"",
"",
"",
"Peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"leng ",
"",
"",
"",
"",
"",
"dan ",
"",
"",
"Xuong ",
"Xuong ",
"",
"",
"du ",
"bian ",
"",
"qia ",
"he ",
"",
"",
"",
"yan ",
"",
"",
"",
"",
"teng ",
"",
"",
"",
"Hom ",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jue ",
"",
"",
"",
"pu ",
"",
"",
"Sun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"So ",
"",
"",
"",
"",
"",
"",
"",
"Cut ",
"",
"",
"kai ",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"sao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
],
"x29b": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zam ",
"",
"",
"Chom ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"zha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xui ",
"",
"",
"dao ",
"",
"",
"an ",
"",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quan ",
"Rau ",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Quan ",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ria ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Rau ",
"",
"",
"zun ",
"",
"nao ",
"",
"",
"",
"",
"",
"",
"",
"cheng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"yao ",
"",
"",
"can ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ban ",
"",
"Xoam ",
"",
"pu ",
"zhuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
],
"x29c": [
"",
"",
"",
"zhan ",
"",
"",
"",
"",
"Nheo ",
"",
"",
"",
"",
"bian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"jue ",
"",
"yu ",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"xie ",
"er ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"er ,xian",
"",
"yu ",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ku ",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hao ",
"",
"",
"niu ",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"Coi ",
"",
"",
"",
"Troi ",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"lai ",
"",
"",
"",
"Ranh ",
"",
"wu ",
"",
"fu ",
"zhuan ",
"",
"",
"su ",
"",
"yao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yan ",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Troi ",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x29d": [
"",
"",
"",
"",
"",
"",
"",
"shai ",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"jiang ",
"",
"",
"",
"",
"",
"ju ",
"",
"Troi ",
"",
"",
"teng ",
"wei ",
"",
"",
"gu ",
"",
"",
"liao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"meng ",
"cha ",
"",
"",
"",
"",
"",
"qu ",
"",
"lei ",
"",
"",
"",
"",
"",
"",
"",
"Troi ",
"",
"qu ",
"",
"",
"",
"",
"",
"Gyo ",
"",
"qiu ",
"",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"du ",
"",
"",
"Chai ",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"Tom ",
"",
"",
"",
"",
"zhu ",
"",
"zhu ",
"Ci ",
"",
"",
"",
"Pou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mu ",
"",
"",
"",
"",
"",
"",
"",
"",
"Buop ",
"",
"meng ",
"",
"",
"",
"",
"guai ",
"jiu ",
"",
"mu ",
"",
"",
"Si ",
"wu ",
"",
"ru ",
"",
"zha ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"jiang ",
"",
"",
"Hung ",
"",
"",
"Thu ",
"",
"",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"Luon ",
"bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiang ",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"Mang ",
"",
"",
"",
"",
"qia ",
"",
"",
"",
"",
],
"x29e": [
"",
"",
"",
"tu ",
"hua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ru ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Yi ",
"",
"",
"",
"",
"",
"Bong ",
"Chuoi ",
"Go ",
"",
"",
"",
"",
"",
"",
"",
"Thia ",
"",
"",
"",
"",
"",
"ye ",
"",
"",
"",
"",
"",
"hai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"an ",
"ba ",
"",
"han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"nai ",
"",
"",
"",
"",
"Thu ",
"",
"",
"",
"",
"Nau ",
"",
"Tep ",
"",
"Chay ",
"Mu ",
"Tuoi ",
"",
"chu ",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"han ",
"",
"na ",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bon ",
"",
"Trau ",
"",
"",
"",
"",
"Bong ",
"",
"Bon ",
"xie ",
"",
"",
"",
"Soc ",
"yu ,wu",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Mam ",
"",
"Sop ",
"Dua ",
"",
"Tram ",
"",
"bu ",
"jian ",
"",
"wu ",
"Tram ",
"",
"zhuan ",
"",
"Vay ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"",
"",
"",
"",
"Vay ",
"Leo ",
"",
"",
"",
"",
],
"x29f": [
"Vay ",
"",
"",
"",
"",
"",
"",
"",
"Ao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ge ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"Sau ",
"Lui ",
"",
"",
"",
"wei ",
"",
"",
"",
"",
"",
"",
"",
"Ruoc ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sau ",
"Luon ",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ba ",
"Nheo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"han ,yan",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"fu ",
"",
"",
"mao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"bao ",
"",
"",
"",
"",
"Mao ",
"",
"Khuou ",
"",
"",
"ju ",
"Caa ",
"qu ",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"Khuou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a0": [
"",
"",
"",
"",
"Chim ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Shi ",
"",
"",
"",
"",
"",
"jiang ",
"",
"",
"",
"",
"",
"",
"",
"xun ",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"jie ",
"",
"yu ",
"",
"",
"",
"Ket ",
"Sac ",
"Hau ",
"Song ",
"zhuang ",
"",
"",
"cheng ",
"",
"jie ",
"",
"chen ",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"yan ",
"",
"",
"",
"ju ",
"",
"",
"",
"",
"",
"",
"",
"",
"xiu ",
"",
"",
"",
"",
"",
"Coc ",
"Coc ",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"fu ",
"ta ",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cun ",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"pen ",
"fu ",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"yu ",
"yu ",
"Mei ",
"",
"mao ",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Hau ",
"",
"",
"Ga ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ge ",
],
"x2a1": [
"",
"",
"xia ",
"",
"",
"Set ",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"ge ",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cut ",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ,gu",
"",
"",
"",
"",
"mai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cuoc ",
"Choi ",
"Ri ",
"Sam ",
"",
"",
"",
"Tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"qu ",
"",
"",
"",
"",
"",
"",
"",
"",
"xie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"Khuou ",
"Tu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ya ",
"liu ",
"",
"",
"",
"",
"can ",
"64272.110,chu ",
"",
"",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"chu ,du",
"",
"ai ",
"",
"",
"Cui ",
"",
"",
"Quam ",
"",
"",
"xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Ngan ",
"",
"Cui ",
"",
"cai ",
"",
"liu ",
"",
"",
"jie ",
"",
"",
"",
"",
"",
"",
"la ",
"",
"",
"",
"",
"",
"",
],
"x2a2": [
"",
"",
"",
"",
"",
"",
"",
"",
"lai ",
"",
"he ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"guan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Muoi ",
"",
"tan ",
"",
"En ",
"",
"",
"",
"",
"",
"kan ",
"",
"bian ",
"",
"",
"Muoi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"gan ",
"",
"",
"",
"",
"",
"Mam ",
"",
"gan ,tan",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"zhu ",
"",
"yu ",
"",
"",
"",
"",
"",
"jian ",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"zu ",
"",
"",
"nuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a3": [
"",
"",
"",
"",
"",
"na ",
"qu ",
"",
"",
"",
"",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"ku ",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"ze ",
"",
"",
"",
"ge ",
"",
"",
"",
"jie ",
"",
"",
"tiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"hun ",
"",
"nie ",
"",
"jun ",
"hu ",
"",
"lu ",
"",
"",
"",
"chao ",
"",
"",
"",
"",
"",
"",
"fu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xuan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"shan ",
"",
"Qu ",
"du ",
"",
"sao ",
"",
"",
"",
"kuang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"liu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mei ",
"",
"",
"",
"",
"",
"",
"tun ",
"kang ",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"heng ",
"",
"",
"",
"",
"",
"Huang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"kuang ",
"piao ",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"bao ",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
],
"x2a4": [
"Naa ",
"",
"",
"",
"",
"",
"bie ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dai ,tai",
"du ",
"",
"",
"",
"",
"",
"",
"tai ",
"",
"shu ",
"",
"",
"",
"",
"",
"",
"su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cha ",
"",
"lu ",
"",
"yu ",
"",
"yan ",
"",
"qiao ",
"",
"yu ",
"",
"",
"tu ",
"",
"Ngam ",
"tun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"die ",
"cha ",
"dian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"wai ",
"",
"",
"",
"",
"",
"zhai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"lu ",
"",
"",
"",
"",
"",
"",
"",
"Tham ",
"",
"",
"",
"",
"ma ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"mai ",
"",
"Nung ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"dan ",
"teng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"yu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Sam ",
"",
"",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a5": [
"",
"",
"Ding ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"peng ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"cu ",
"jun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"er ",
"",
"",
"",
"",
"ai ",
"hu ",
"",
"",
"hu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jiao ",
"",
"",
"",
"pu ",
"",
"",
"jie ",
"lu ",
"",
"",
"",
"yao ,ya",
"",
"",
"",
"",
"",
"",
"",
"",
"Hou ",
"",
"qiu ",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"xu ",
"",
"",
"",
"Ngui ",
"",
"",
"",
"",
"",
"",
"",
"",
"su ",
"liao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"pa ",
"",
"",
"",
"",
"",
"",
"",
"",
"na ",
"",
"",
"",
"",
"",
"",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a6": [
"",
"Kap6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zu ",
"",
"",
"zhan ",
"",
"",
"",
"",
"",
"",
"",
"Ji ",
"",
"",
"Rang ",
"",
"",
"",
"he ",
"qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"hu ",
"",
"yan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nak ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Trong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhe ",
"",
"",
"",
"Gwi ",
"",
"",
"gan ",
"",
"",
"cu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"zhu ",
"",
"",
"",
"",
"",
"",
"xiao ",
"",
"",
"",
"",
"Rua ",
"",
"Rua ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a7": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2a8": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Bai ",
"",
"",
"",
"",
"",
"",
"Zhan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Luan ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2aa": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Song ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Yong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2ae": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Nu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Cong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2af": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xian ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b0": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Li ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Fei ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Su ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Kou ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b1": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chi ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xun ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b2": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Qia ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b3": [
"Ji ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Luo ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Yi ",
"",
"",
"",
"",
"",
"",
"",
"",
"Nao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xi ",
"",
"Xiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jiao ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b4": [
"",
"",
"",
"",
"Yue ",
"",
"Kuai ",
"",
"",
"Ling ",
"",
"",
"",
"",
"",
"",
"Ni ",
"",
"",
"Bu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Han ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Fu ",
"",
"Cong ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b5": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Jue ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Duo ",
"",
"",
"",
"",
"",
"",
"Su ",
"",
"",
"",
"",
"",
"",
"Huang ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2b6": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Han ",
"Ai ",
"",
"",
"",
"Ti ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Xu ",
"Gong ",
"",
"",
"",
"",
"",
"",
"",
"",
"Ping ",
"",
"Hui ",
"Shi ",
"",
"",
"",
"Pu ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Zhi ",
"",
"",
"Jue ",
"",
"",
"",
"Ning ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Chi ",
"",
"Ti ",
"",
"",
"",
"",
"",
"",
"",
],
"x2f8": [
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"Gai ",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
"x2f9": [
"",
"",
"",
"",
"",
"",
"",
"Baan6",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
"",
],
}
|
updater | zeroname_updater | from __future__ import print_function
import json
import os
import re
import socket
import sys
import time
from subprocess import call
from bitcoinrpc.authproxy import AuthServiceProxy
from six import string_types
def publish():
print("* Signing and Publishing...")
call(" ".join(command_sign_publish), shell=True)
def processNameOp(domain, value, test=False):
if not value.strip().startswith("{"):
return False
try:
data = json.loads(value)
except Exception as err:
print("Json load error: %s" % err)
return False
if "zeronet" not in data and "map" not in data:
# Namecoin standard use {"map": { "blog": {"zeronet": "1D..."} }}
print("No zeronet and no map in ", data.keys())
return False
if "map" in data:
# If subdomains using the Namecoin standard is present, just re-write in the Zeronet way
# and call the function again
data_map = data["map"]
new_value = {}
for subdomain in data_map:
if "zeronet" in data_map[subdomain]:
new_value[subdomain] = data_map[subdomain]["zeronet"]
if "zeronet" in data and isinstance(data["zeronet"], string_types):
# {
# "zeronet":"19rXKeKptSdQ9qt7omwN82smehzTuuq6S9",
# ....
# }
new_value[""] = data["zeronet"]
if len(new_value) > 0:
return processNameOp(domain, json.dumps({"zeronet": new_value}), test)
else:
return False
if "zeronet" in data and isinstance(data["zeronet"], string_types):
# {
# "zeronet":"19rXKeKptSdQ9qt7omwN82smehzTuuq6S9"
# } is valid
return processNameOp(
domain, json.dumps({"zeronet": {"": data["zeronet"]}}), test
)
if not isinstance(data["zeronet"], dict):
print("Not dict: ", data["zeronet"])
return False
if not re.match("^[a-z0-9]([a-z0-9-]{0,62}[a-z0-9])?$", domain):
print("Invalid domain: ", domain)
return False
if test:
return True
if "slave" in sys.argv:
print("Waiting for master update arrive")
time.sleep(30) # Wait 30 sec to allow master updater
# Note: Requires the file data/names.json to exist and contain "{}" to work
names_raw = open(names_path, "rb").read()
names = json.loads(names_raw)
for subdomain, address in data["zeronet"].items():
subdomain = subdomain.lower()
address = re.sub("[^A-Za-z0-9]", "", address)
print(subdomain, domain, "->", address)
if subdomain:
if re.match("^[a-z0-9]([a-z0-9-]{0,62}[a-z0-9])?$", subdomain):
names["%s.%s.bit" % (subdomain, domain)] = address
else:
print("Invalid subdomain:", domain, subdomain)
else:
names["%s.bit" % domain] = address
new_names_raw = json.dumps(names, indent=2, sort_keys=True)
if new_names_raw != names_raw:
open(names_path, "wb").write(new_names_raw)
print("-", domain, "Changed")
return True
else:
print("-", domain, "Not changed")
return False
def processBlock(block_id, test=False):
print("Processing block #%s..." % block_id)
s = time.time()
block_hash = rpc.getblockhash(block_id)
block = rpc.getblock(block_hash)
print("Checking %s tx" % len(block["tx"]))
updated = 0
for tx in block["tx"]:
try:
transaction = rpc.getrawtransaction(tx, 1)
for vout in transaction.get("vout", []):
if (
"scriptPubKey" in vout
and "nameOp" in vout["scriptPubKey"]
and "name" in vout["scriptPubKey"]["nameOp"]
):
name_op = vout["scriptPubKey"]["nameOp"]
updated += processNameOp(
name_op["name"].replace("d/", ""), name_op["value"], test
)
except Exception as err:
print("Error processing tx #%s %s" % (tx, err))
print("Done in %.3fs (updated %s)." % (time.time() - s, updated))
return updated
# Connecting to RPC
def initRpc(config):
"""Initialize Namecoin RPC"""
rpc_data = {
"connect": "127.0.0.1",
"port": "8336",
"user": "PLACEHOLDER",
"password": "PLACEHOLDER",
"clienttimeout": "900",
}
try:
fptr = open(config, "r")
lines = fptr.readlines()
fptr.close()
except:
return None # Or take some other appropriate action
for line in lines:
if not line.startswith("rpc"):
continue
key_val = line.split(None, 1)[0]
(key, val) = key_val.split("=", 1)
if not key or not val:
continue
rpc_data[key[3:]] = val
url = "http://%(user)s:%(password)s@%(connect)s:%(port)s" % rpc_data
return url, int(rpc_data["clienttimeout"])
# Loading config...
# Check whether platform is on windows or linux
# On linux namecoin is installed under ~/.namecoin, while on on windows it is in %appdata%/Namecoin
if sys.platform == "win32":
namecoin_location = os.getenv("APPDATA") + "/Namecoin/"
else:
namecoin_location = os.path.expanduser("~/.namecoin/")
config_path = namecoin_location + "zeroname_config.json"
if not os.path.isfile(config_path): # Create sample config
open(config_path, "w").write(
json.dumps(
{
"site": "site",
"zeronet_path": "/home/zeronet",
"privatekey": "",
"lastprocessed": 223910,
},
indent=2,
)
)
print("* Example config written to %s" % config_path)
sys.exit(0)
config = json.load(open(config_path))
names_path = "%s/data/%s/data/names.json" % (config["zeronet_path"], config["site"])
os.chdir(
config["zeronet_path"]
) # Change working dir - tells script where Zeronet install is.
# Parameters to sign and publish
command_sign_publish = [
sys.executable,
"zeronet.py",
"siteSign",
config["site"],
config["privatekey"],
"--publish",
]
if sys.platform == "win32":
command_sign_publish = ['"%s"' % param for param in command_sign_publish]
# Initialize rpc connection
rpc_auth, rpc_timeout = initRpc(namecoin_location + "namecoin.conf")
rpc = AuthServiceProxy(rpc_auth, timeout=rpc_timeout)
node_version = rpc.getnetworkinfo()["version"]
while 1:
try:
time.sleep(1)
if node_version < 160000:
last_block = int(rpc.getinfo()["blocks"])
else:
last_block = int(rpc.getblockchaininfo()["blocks"])
break # Connection succeeded
except socket.timeout: # Timeout
print(".", end=" ")
sys.stdout.flush()
except Exception as err:
print("Exception", err.__class__, err)
time.sleep(5)
rpc = AuthServiceProxy(rpc_auth, timeout=rpc_timeout)
if not config["lastprocessed"]: # First startup: Start processing from last block
config["lastprocessed"] = last_block
print("- Testing domain parsing...")
assert processBlock(223911, test=True) # Testing zeronetwork.bit
assert processBlock(227052, test=True) # Testing brainwallets.bit
assert not processBlock(236824, test=True) # Utf8 domain name (invalid should skip)
assert not processBlock(236752, test=True) # Uppercase domain (invalid should skip)
assert processBlock(236870, test=True) # Encoded domain (should pass)
assert processBlock(
438317, test=True
) # Testing namecoin standard artifaxradio.bit (should pass)
# sys.exit(0)
print("- Parsing skipped blocks...")
should_publish = False
for block_id in range(config["lastprocessed"], last_block + 1):
if processBlock(block_id):
should_publish = True
config["lastprocessed"] = last_block
if should_publish:
publish()
while 1:
print("- Waiting for new block")
sys.stdout.flush()
while 1:
try:
time.sleep(1)
if node_version < 160000:
rpc.waitforblock()
else:
rpc.waitfornewblock()
print("Found")
break # Block found
except socket.timeout: # Timeout
print(".", end=" ")
sys.stdout.flush()
except Exception as err:
print("Exception", err.__class__, err)
time.sleep(5)
rpc = AuthServiceProxy(rpc_auth, timeout=rpc_timeout)
if node_version < 160000:
last_block = int(rpc.getinfo()["blocks"])
else:
last_block = int(rpc.getblockchaininfo()["blocks"])
should_publish = False
for block_id in range(config["lastprocessed"] + 1, last_block + 1):
if processBlock(block_id):
should_publish = True
config["lastprocessed"] = last_block
open(config_path, "w").write(json.dumps(config, indent=2))
if should_publish:
publish()
|
neubot | marshal | # neubot/marshal.py
# -*- coding: utf-8 -*-
#
# Copyright (c) 2011 Simone Basso <bassosimone@gmail.com>,
# NEXA Center for Internet & Society at Politecnico di Torino
#
# This file is part of Neubot <http://www.neubot.org/>.
#
# Neubot is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Neubot is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Neubot. If not, see <http://www.gnu.org/licenses/>.
#
import cgi
import sys
import types
import urllib
import xml.dom.minidom
if __name__ == "__main__":
sys.path.insert(0, ".")
from neubot.compat import json
from neubot.utils import stringify, unicodize
# Marshal
def object_to_json(obj):
return json.dumps(obj.__dict__)
def object_to_qs(obj):
dictionary = {}
for name, value in obj.__dict__.items():
dictionary[name] = stringify(value)
s = urllib.urlencode(dictionary)
return s
def object_to_xml(obj):
return dict_to_xml(obj.__class__.__name__, obj.__dict__)
def dict_to_xml(name, thedict):
document = xml.dom.minidom.parseString("<" + name + "/>")
root = document.documentElement
root.appendChild( document.createTextNode("\r\n") )
for name, value in thedict.items():
if type(value) != types.ListType:
value = [value]
for v in value:
root.appendChild( document.createTextNode(" ") )
element = document.createElement(name)
root.appendChild(element)
element.appendChild( document.createTextNode( unicodize(v) ))
root.appendChild( document.createTextNode("\r\n") )
return root.toxml("utf-8")
MARSHALLERS = {
"application/json": object_to_json,
"application/x-www-form-urlencoded": object_to_qs,
"application/xml": object_to_xml,
"text/xml": object_to_xml,
}
def marshal_object(obj, mimetype):
return MARSHALLERS[mimetype](obj)
# Unmarshal
def json_to_dictionary(s):
return dict(json.loads(s))
def qs_to_dictionary(s):
dictionary = {}
for name, value in cgi.parse_qs(s).items():
dictionary[name] = value[0]
return dictionary
def xml_to_dictionary(s):
dictionary = {}
document = xml.dom.minidom.parseString(s)
document.documentElement.normalize() # XXX
for element in document.documentElement.childNodes:
if element.nodeType == element.ELEMENT_NODE:
for node in element.childNodes:
if node.nodeType == node.TEXT_NODE:
if not element.tagName in dictionary:
dictionary[element.tagName] = []
dictionary[element.tagName].append(node.data.strip())
break
return dictionary
UNMARSHALLERS = {
"application/json": json_to_dictionary,
"application/x-www-form-urlencoded": qs_to_dictionary,
"application/xml": xml_to_dictionary,
"text/xml": xml_to_dictionary,
}
def unmarshal_objectx(s, mimetype, instance):
dictionary = UNMARSHALLERS[mimetype](s)
for name, value in instance.__dict__.items():
if name in dictionary:
nval = dictionary[name]
if type(value) != types.ListType and type(nval) == types.ListType:
nval = nval[0]
setattr(instance, name, nval)
def unmarshal_object(s, mimetype, ctor):
obj = ctor()
unmarshal_objectx(s, mimetype, obj)
return obj
# Unit test
import pprint
class Test(object):
def __init__(self):
# the city name that prompted all unicode issues...
self.uname = u"Aglié"
self.sname = "Aglie"
self.fval = 1.43
self.ival = 1<<17
self.vect = [1,2,3,4,5,12,u"Aglié",1.43]
self.v2 = ["urbinek"]
def test(mimetype):
pprint.pprint("--- %s ---" % mimetype)
m = Test()
pprint.pprint(m.__dict__)
e = marshal_object(m, mimetype)
print e
d = unmarshal_object(e, mimetype, Test)
pprint.pprint(d.__dict__)
if __name__ == "__main__":
if len(sys.argv) == 1:
mimetypes = MARSHALLERS.keys()
else:
mimetypes = sys.argv[1:]
for mimetype in mimetypes:
test(mimetype)
|
frescobaldi-app | actioncollection | # This file is part of the Frescobaldi project, http://www.frescobaldi.org/
#
# Copyright (c) 2008 - 2014 by Wilbert Berendsen
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
# See http://www.gnu.org/licenses/ for more information.
"""
In this module are two classes, ActionCollection and ShortcutCollection.
Both must be inherited to do something useful.
ActionCollection keeps a fixed list of QActions, set as instance attributes in
the createActions() method. The icons and default shortcuts may also be set in
the same method. The texts should be set in the translateUI() method.
The ActionCollection then keeps track of possibly changed keyboard shortcuts by
loading them from the config and connecting to the app.settingsChanged() signal.
ShortcutCollection keeps a variable list of QActions, for which default
shortcuts must be set in the createDefaultShortcuts() method.
This actions must not be connected to, but they are only used to set keyboard
shortcuts for a module that needs not to be loaded initially for the shortcuts
to work. If a shortcut is pressed, the real action is queried using the
realAction() method, which should return the corresponding action in the UI.
That one is than triggered.
The module may provide the user with a means to change the keyboard shortcuts,
which then should call setShortcuts() to do it. The module may also query the
currently set shortcuts for an action using shortcuts().
"""
import weakref
import app
from PyQt5.QtCore import QSettings, Qt
from PyQt5.QtGui import QKeySequence
from PyQt5.QtWidgets import QAction
class ActionCollectionBase:
"""Abstract base class. Can load and keep a list of QActions.
You must subclass this class and provide a name for the actioncollection
in the 'name' class attribute.
"""
def __init__(self, widget=None):
self._widget = weakref.ref(widget) if widget else lambda: None
self._actions = {} # maps name to action
self._defaults = {} # maps name to default list of shortcuts
app.settingsChanged.connect(self.load)
def widget(self):
"""Returns the widget given on construction or None."""
return self._widget()
def setDefaultShortcuts(self, name, shortcuts):
"""Set a default list of QKeySequence objects for the named action."""
self._defaults[name] = shortcuts
def defaultShortcuts(self, name):
"""Returns the default shortcuts (list of QKeySequences) for the action.
If not defined, returns None.
"""
return self._defaults.get(name)
def actions(self):
"""Returns the dictionary with actions."""
return self._actions
def defaults(self):
"""Returns the dictionary with actions that have a default shortcut."""
return self._defaults
def shortcuts(self, name):
"""Returns the list of shortcuts for the named action, or None."""
try:
return self._actions[name].shortcuts()
except KeyError:
pass
def setShortcuts(self, name, shortcuts):
"""Implement to set the shortcuts list for our action."""
pass
def settingsGroup(self):
"""Returns settings group to load/save shortcuts from or to."""
s = QSettings()
scheme = s.value("shortcut_scheme", "default", str)
s.beginGroup(f"shortcuts/{scheme}/{self.name}")
return s
def load(self):
"""Implement to load shortcuts from our settingsGroup()."""
pass
def title(self):
"""If this returns a meaningful title, actions can be grouped in the shortcut settings dialog."""
pass
class ActionCollection(ActionCollectionBase):
"""Keeps a fixed list of QActions as instance attributes.
Subclass this and add the actions as instance attributes in
the createActions() method.
You can set the default shortcuts directly in the actions in the
createActions() method, it is not needed to use the setDefaultShortcuts()
method for that.
Set the titles for the actions in the translateUI() method.
"""
def __init__(self, parent=None):
"""Creates the ActionCollection.
parent is an optional widget that is also the parent for the created actions.
"""
super().__init__(parent)
self.createActions(parent)
self._actions = dict(
i for i in self.__dict__.items() if not i[0].startswith("_")
)
self.storeDefaults()
self.load(False) # load the shortcuts without resettings defaults
# N.B. This relies on the ActionCollection's QActions not getting
# deleted in C++ other than through being garbage-collected in Python.
app.translateUI(self)
def createActions(self, parent=None):
"""Should add actions as instance attributes.
The QActions should get icons and shortcuts. Texts should be set
in translateUI(). The actions are created with the parent given on instantiation.
"""
pass
def translateUI(self):
"""Should (re)translate all the titles of the actions."""
pass
def storeDefaults(self):
"""Saves the preset default QKeySequence lists for the actions."""
for name, action in self._actions.items():
if action.shortcuts():
self.setDefaultShortcuts(name, action.shortcuts())
def setShortcuts(self, name, shortcuts):
"""Sets the shortcuts list for our action. Use an empty list to remove the shortcuts."""
action = self.actions().get(name)
if not action:
return
default = self.defaultShortcuts(name)
setting = self.settingsGroup()
action.setShortcuts(shortcuts)
setting.setValue(name, shortcuts)
if default:
if shortcuts == default:
setting.remove(name)
else:
setting.setValue(name, shortcuts)
else:
if shortcuts:
setting.setValue(name, shortcuts)
else:
setting.remove(name)
def load(self, restoreDefaults=True):
"""Reads keyboard shortcuts from the settings.
If restoreDefaults == True, resets the other shortcuts to their default
values. If restoreDefaults == False, does not touch the other shortcuts.
"""
settings = self.settingsGroup()
keys = settings.allKeys()
for name in keys:
try:
shortcuts = settings.value(name, [], QKeySequence)
except TypeError:
# PyQt5 raises TypeError when an empty list was stored
shortcuts = []
try:
self._actions[name].setShortcuts(shortcuts)
except KeyError:
settings.remove(name)
if restoreDefaults:
for name in self._actions:
if name not in keys:
self._actions[name].setShortcuts(self._defaults.get(name) or [])
class ShortcutCollection(ActionCollectionBase):
"""An ActionCollection type that only saves actions that have a keyboard shortcut.
Should always be instantiated with a visible widget (preferably MainWindow)
as parent.
Use the setShortcuts() method to set a list (possibly empty) of QKeySequence
objects. Every change causes other instances of the same-named collection
to reload.
This serves two purposes:
1. save keyboard shortcuts for actions created by the user or from a very large list
2. make the keyboard shortcuts working even if the component the actions are
contained in is not even loaded yet.
To make this work, implement the realAction() method to instantiate the widget the
action is meant for and then return the real action.
"""
# save weak references to other instances with the same name and sync then.
others = {}
# shortcut context to use by default
shortcutContext = Qt.WindowShortcut
def __init__(self, widget):
"""Creates the ShortcutCollection.
The widget is required as actions are added to it, so their keyboard
shortcuts get triggered.
"""
super().__init__(widget)
self.createDefaultShortcuts()
self.load()
self.others.setdefault(self.name, []).append(weakref.ref(self))
def createDefaultShortcuts(self):
"""Should set some default shortcut lists using setDefaultShortcuts()."""
pass
def load(self):
"""Reads keyboard shortcuts from the settings. Instantiates QActions as needed."""
# clears all actions
for a in self._actions.values():
a.setParent(None)
self._actions = {}
# then set defaults
for name, shortcuts in self.defaults().items():
self.action(name).setShortcuts(shortcuts)
# then load
settings = self.settingsGroup()
for name in settings.allKeys():
try:
shortcuts = settings.value(name, [], QKeySequence)
except TypeError:
# PyQt5 raises TypeError when an empty list was stored
shortcuts = []
if not shortcuts:
if not self.removeAction(name):
# if it did not exist, remove key from config
settings.remove(name)
else:
self.action(name).setShortcuts(shortcuts)
def setShortcuts(self, name, shortcuts):
"""Sets the shortcuts list for our action. Use an empty list to remove the shortcuts."""
if shortcuts:
self.action(name).setShortcuts(shortcuts)
self.settingsGroup().setValue(name, shortcuts)
else:
self.removeAction(name)
if name in self.defaults():
# save empty list because there exists a default value
self.settingsGroup().setValue(name, [])
else:
self.settingsGroup().remove(name)
self.reloadOthers()
def restoreDefaultShortcuts(self, name):
"""Resets the shortcuts for the specified action to their default value."""
shortcuts = self.defaultShortcuts(name)
if shortcuts:
self.action(name).setShortcuts(shortcuts)
else:
self.removeAction(name)
self.settingsGroup().remove(name)
self.reloadOthers()
def removeAction(self, name):
"""(Internal) Removes the named action, returning True it it did exist."""
try:
a = self._actions[name]
except KeyError:
return False
a.setParent(None)
del self._actions[name]
return True
def action(self, name):
"""Returns a QAction for the name, instantiating it if necessary."""
try:
a = self._actions[name]
except KeyError:
a = self._actions[name] = QAction(self.widget())
a.setShortcutContext(self.shortcutContext)
a.triggered.connect(lambda: self.triggerAction(name))
self.widget().addAction(a)
return a
def triggerAction(self, name):
"""Called when the user presses a saved keyboard shortcut."""
a = self.realAction(name)
if a:
a.trigger()
def realAction(self, name):
"""Implement this to return the real action the name refers to,
This is called when the text and icon are needed (e.g. when the shortcut
dialog is opened) or when our "shadow" action keyboard shortcut is triggered.
The function may return None, e.g. when the action our name refers to does
not exist anymore. In that case our action is also removed.
"""
pass
def actions(self):
"""Returns our real actions instead of the shadow ones."""
d = {}
changed = False
for name in list(self._actions):
a = self.realAction(name)
if a:
d[name] = a
else:
self.removeAction(name)
changed = True
if changed:
self.reloadOthers()
return d
def reloadOthers(self):
"""Reload others managing the same shortcuts (e.g. in case of multiple mainwindows)."""
for ref in self.others[self.name][:]:
other = ref()
if not other:
self.others[self.name].remove(ref)
elif other is not self:
other.load()
|
management | forms | # -*- coding: utf-8 -*-
"""
flaskbb.management.forms
~~~~~~~~~~~~~~~~~~~~~~~~
It provides the forms that are needed for the management views.
:copyright: (c) 2014 by the FlaskBB Team.
:license: BSD, see LICENSE for more details.
"""
import logging
from flask_allows import Permission
from flask_babelplus import lazy_gettext as _
from flask_wtf import FlaskForm
from flaskbb.extensions import db
from flaskbb.forum.models import Category, Forum
from flaskbb.user.models import Group, User
from flaskbb.utils.helpers import check_image
from flaskbb.utils.requirements import IsAtleastModerator
from sqlalchemy.orm.session import make_transient, make_transient_to_detached
from wtforms import (
BooleanField,
DateField,
HiddenField,
IntegerField,
PasswordField,
StringField,
SubmitField,
TextAreaField,
)
from wtforms.validators import (
URL,
DataRequired,
Email,
Length,
Optional,
ValidationError,
regexp,
)
from wtforms_sqlalchemy.fields import QuerySelectField, QuerySelectMultipleField
logger = logging.getLogger(__name__)
USERNAME_RE = r"^[\w.+-]+$"
is_username = regexp(
USERNAME_RE, message=_("You can only use letters, numbers or dashes.")
)
def selectable_forums():
return Forum.query.order_by(Forum.position)
def selectable_categories():
return Category.query.order_by(Category.position)
def selectable_groups():
return Group.query.order_by(Group.id.asc()).all()
def select_primary_group():
return Group.query.filter(Group.guest != True).order_by(Group.id)
class UserForm(FlaskForm):
username = StringField(
_("Username"),
validators=[
DataRequired(message=_("A valid username is required.")),
is_username,
],
)
email = StringField(
_("Email address"),
validators=[
DataRequired(message=_("A valid email address is required.")),
Email(message=_("Invalid email address.")),
],
)
password = PasswordField("Password", validators=[DataRequired()])
birthday = DateField(_("Birthday"), format="%Y-%m-%d", validators=[Optional()])
gender = StringField(_("Gender"), validators=[Optional()])
location = StringField(_("Location"), validators=[Optional()])
website = StringField(_("Website"), validators=[Optional(), URL()])
avatar = StringField(_("Avatar"), validators=[Optional(), URL()])
signature = TextAreaField(_("Forum signature"), validators=[Optional()])
notes = TextAreaField(_("Notes"), validators=[Optional(), Length(min=0, max=5000)])
activated = BooleanField(_("Is active?"), validators=[Optional()])
primary_group = QuerySelectField(
_("Primary group"), query_factory=select_primary_group, get_label="name"
)
secondary_groups = QuerySelectMultipleField(
_("Secondary groups"),
# TODO: Template rendering errors "NoneType is not callable"
# without this, figure out why.
query_factory=select_primary_group,
get_label="name",
)
submit = SubmitField(_("Save"))
def validate_username(self, field):
if hasattr(self, "user"):
user = User.query.filter(
db.and_(
User.username.like(field.data.lower()),
db.not_(User.id == self.user.id),
)
).first()
else:
user = User.query.filter(User.username.like(field.data.lower())).first()
if user:
raise ValidationError(_("This username is already taken."))
def validate_email(self, field):
if hasattr(self, "user"):
user = User.query.filter(
db.and_(
User.email.like(field.data.lower()),
db.not_(User.id == self.user.id),
)
).first()
else:
user = User.query.filter(User.email.like(field.data.lower())).first()
if user:
raise ValidationError(_("This email address is already taken."))
def validate_avatar(self, field):
if field.data is not None:
error, status = check_image(field.data)
if error is not None:
raise ValidationError(error)
return status
def save(self):
data = self.data
data.pop("submit", None)
data.pop("csrf_token", None)
user = User(**data)
return user.save()
class AddUserForm(UserForm):
pass
class EditUserForm(UserForm):
password = PasswordField("Password", validators=[Optional()])
def __init__(self, user, *args, **kwargs):
self.user = user
kwargs["obj"] = self.user
UserForm.__init__(self, *args, **kwargs)
class GroupForm(FlaskForm):
name = StringField(
_("Group name"),
validators=[DataRequired(message=_("Please enter a name for the group."))],
)
description = TextAreaField(_("Description"), validators=[Optional()])
admin = BooleanField(
_("Is 'Admin' group?"),
description=_("With this option the group has access to " "the admin panel."),
)
super_mod = BooleanField(
_("Is 'Super Moderator' group?"),
description=_(
"Check this, if the users in this group are allowed to "
"moderate every forum."
),
)
mod = BooleanField(
_("Is 'Moderator' group?"),
description=_(
"Check this, if the users in this group are allowed to "
"moderate specified forums."
),
)
banned = BooleanField(
_("Is 'Banned' group?"),
description=_("Only one group of type 'Banned' is allowed."),
)
guest = BooleanField(
_("Is 'Guest' group?"),
description=_("Only one group of type 'Guest' is allowed."),
)
editpost = BooleanField(
_("Can edit posts"),
description=_("Check this, if the users in this group can edit posts."),
)
deletepost = BooleanField(
_("Can delete posts"),
description=_("Check this, if the users in this group can delete " "posts."),
)
deletetopic = BooleanField(
_("Can delete topics"),
description=_("Check this, if the users in this group can delete " "topics."),
)
posttopic = BooleanField(
_("Can create topics"),
description=_("Check this, if the users in this group can create " "topics."),
)
postreply = BooleanField(
_("Can post replies"),
description=_("Check this, if the users in this group can post " "replies."),
)
mod_edituser = BooleanField(
_("Moderators can edit user profiles"),
description=_(
"Allow moderators to edit another user's profile "
"including password and email changes."
),
)
mod_banuser = BooleanField(
_("Moderators can ban users"),
description=_("Allow moderators to ban other users."),
)
viewhidden = BooleanField(
_("Can view hidden posts and topics"),
description=_("Allows a user to view hidden posts and topics"),
)
makehidden = BooleanField(
_("Can hide posts and topics"),
description=_("Allows a user to hide posts and topics"),
)
submit = SubmitField(_("Save"))
def validate_name(self, field):
if hasattr(self, "group"):
group = Group.query.filter(
db.and_(
Group.name.like(field.data.lower()),
db.not_(Group.id == self.group.id),
)
).first()
else:
group = Group.query.filter(Group.name.like(field.data.lower())).first()
if group:
raise ValidationError(_("This group name is already taken."))
def validate_banned(self, field):
if hasattr(self, "group"):
group = Group.query.filter(
db.and_(Group.banned, db.not_(Group.id == self.group.id))
).count()
else:
group = Group.query.filter_by(banned=True).count()
if field.data and group > 0:
raise ValidationError(_("There is already a group of type " "'Banned'."))
def validate_guest(self, field):
if hasattr(self, "group"):
group = Group.query.filter(
db.and_(Group.guest, db.not_(Group.id == self.group.id))
).count()
else:
group = Group.query.filter_by(guest=True).count()
if field.data and group > 0:
raise ValidationError(_("There is already a group of type " "'Guest'."))
def validate(self):
if not super(GroupForm, self).validate():
return False
result = True
permission_fields = (
self.editpost,
self.deletepost,
self.deletetopic,
self.posttopic,
self.postreply,
self.mod_edituser,
self.mod_banuser,
self.viewhidden,
self.makehidden,
)
group_fields = [self.admin, self.super_mod, self.mod, self.banned, self.guest]
# we do not allow to modify any guest permissions
if self.guest.data:
for field in permission_fields:
if field.data:
# if done in 'validate_guest' it would display this
# warning on the fields
field.errors.append(
_("Can't assign any permissions to this group.")
)
result = False
checked = []
for field in group_fields:
if field.data and field.data in checked:
if len(checked) > 1:
field.errors.append("A group can't have multiple group types.")
result = False
else:
checked.append(field.data)
return result
def save(self):
data = self.data
data.pop("submit", None)
data.pop("csrf_token", None)
group = Group(**data)
return group.save()
class EditGroupForm(GroupForm):
def __init__(self, group, *args, **kwargs):
self.group = group
kwargs["obj"] = self.group
GroupForm.__init__(self, *args, **kwargs)
class AddGroupForm(GroupForm):
pass
class ForumForm(FlaskForm):
title = StringField(
_("Forum title"),
validators=[DataRequired(message=_("Please enter a forum title."))],
)
description = TextAreaField(
_("Description"),
validators=[Optional()],
description=_("You can format your description with Markdown."),
)
position = IntegerField(
_("Position"),
default=1,
validators=[
DataRequired(message=_("Please enter a position for the" "forum."))
],
)
category = QuerySelectField(
_("Category"),
query_factory=selectable_categories,
allow_blank=False,
get_label="title",
description=_("The category that contains this forum."),
)
external = StringField(
_("External link"),
validators=[Optional(), URL()],
description=_("A link to a website i.e. 'http://flaskbb.org'."),
)
moderators = StringField(
_("Moderators"),
description=_(
"Comma separated usernames. Leave it blank if you do "
"not want to set any moderators."
),
)
show_moderators = BooleanField(
_("Show moderators"),
description=_("Do you want to show the moderators on the index page?"),
)
locked = BooleanField(
_("Locked?"), description=_("Disable new posts and topics in this forum.")
)
groups = QuerySelectMultipleField(
_("Group access"),
query_factory=selectable_groups,
get_label="name",
description=_("Select the groups that can access this forum."),
)
submit = SubmitField(_("Save"))
def validate_external(self, field):
if hasattr(self, "forum"):
if self.forum.topics.count() > 0:
raise ValidationError(
_(
"You cannot convert a forum that "
"contains topics into an "
"external link."
)
)
def validate_show_moderators(self, field):
if field.data and not self.moderators.data:
raise ValidationError(_("You also need to specify some " "moderators."))
def validate_moderators(self, field):
approved_moderators = []
if field.data:
moderators = [mod.strip() for mod in field.data.split(",")]
users = User.query.filter(User.username.in_(moderators))
for user in users:
if not Permission(IsAtleastModerator, identity=user):
raise ValidationError(
_("%(user)s is not in a moderators group.", user=user.username)
)
else:
approved_moderators.append(user)
field.data = approved_moderators
def save(self):
data = self.data
# delete submit and csrf_token from data
data.pop("submit", None)
data.pop("csrf_token", None)
forum = Forum(**data)
return forum.save()
class EditForumForm(ForumForm):
id = HiddenField()
def __init__(self, forum, *args, **kwargs):
self.forum = forum
kwargs["obj"] = self.forum
ForumForm.__init__(self, *args, **kwargs)
def save(self):
data = self.data
# delete submit and csrf_token from data
data.pop("submit", None)
data.pop("csrf_token", None)
forum = Forum(**data)
# flush SQLA info from created instance so that it can be merged
make_transient(forum)
make_transient_to_detached(forum)
return forum.save()
class AddForumForm(ForumForm):
pass
class CategoryForm(FlaskForm):
title = StringField(
_("Category title"),
validators=[DataRequired(message=_("Please enter a category title."))],
)
description = TextAreaField(
_("Description"),
validators=[Optional()],
description=_("You can format your description with Markdown."),
)
position = IntegerField(
_("Position"),
default=1,
validators=[
DataRequired(message=_("Please enter a position for the " "category."))
],
)
submit = SubmitField(_("Save"))
def save(self):
data = self.data
# delete submit and csrf_token from data
data.pop("submit", None)
data.pop("csrf_token", None)
category = Category(**data)
return category.save()
|
plugins | piaulizaportal | """
$description Japanese live-streaming and video hosting platform owned by PIA Corporation.
$url ulizaportal.jp
$type live, vod
$metadata id
$metadata title
$account Purchased tickets are required.
$notes Tickets purchased at "PIA LIVE STREAM" are used for this platform.
"""
import logging
import re
import time
from urllib.parse import parse_qsl, urlparse
from streamlink.plugin import Plugin, pluginmatcher
from streamlink.plugin.api import validate
from streamlink.stream.hls import HLSStream
log = logging.getLogger(__name__)
@pluginmatcher(
re.compile(
r"https://ulizaportal\.jp/pages/(?P<id>[\da-f]{8}-[\da-f]{4}-[\da-f]{4}-[\da-f]{4}-[\da-f]{12})",
)
)
class PIAULIZAPortal(Plugin):
_URL_PLAYER_DATA = "https://player-api.p.uliza.jp/v1/players/"
_URL_PLAYLIST = "https://vms-api.p.uliza.jp/v1/prog-index.m3u8"
def _get_streams(self):
self.id = self.match.group("id")
try:
expires = int(dict(parse_qsl(urlparse(self.url).query)).get("expires", 0))
except ValueError:
expires = 0
if 0 < expires <= time.time():
log.error("The URL has expired")
return None
self.title, player_data_url = self.session.http.get(
self.url,
schema=validate.Schema(
validate.parse_html(),
validate.union(
(
validate.xml_xpath_string(".//head/title[1]/text()"),
validate.xml_xpath_string(
f".//script[@type='text/javascript'][contains(@src,'{self._URL_PLAYER_DATA}')][1]/@src",
),
)
),
),
)
if not player_data_url:
log.error("Player data URL not found")
return None
m3u8_url = self.session.http.get(
player_data_url,
headers={
"Referer": self.url,
},
schema=validate.Schema(
re.compile(rf"""{re.escape(self._URL_PLAYLIST)}[^"']+"""),
validate.none_or_all(
validate.get(0),
validate.url(),
),
),
)
if not m3u8_url:
log.error("Playlist URL not found")
return None
return HLSStream.parse_variant_playlist(self.session, m3u8_url)
__plugin__ = PIAULIZAPortal
|
misc | depcheck | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
# Copyright 2016,2017 Christoph Reiter
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
"""
Deletes unneeded DLLs and checks DLL dependencies.
Execute with the build python, will figure out the rest.
"""
import os
import subprocess
import sys
from multiprocessing import Process, Queue
import gi # isort:skip
gi.require_version("GIRepository", "2.0") # isort:skip
from gi.repository import GIRepository # isort:skip
def _get_shared_libraries(q, namespace, version):
repo = GIRepository.Repository()
repo.require(namespace, version, 0)
lib = repo.get_shared_library(namespace)
q.put(lib)
def get_shared_libraries(namespace, version):
# we have to start a new process because multiple versions can't be loaded
# in the same process
q = Queue()
p = Process(target=_get_shared_libraries, args=(q, namespace, version))
p.start()
result = q.get()
p.join()
return result
def get_required_by_typelibs():
deps = set()
repo = GIRepository.Repository()
for tl in os.listdir(repo.get_search_path()[0]):
namespace, version = os.path.splitext(tl)[0].split("-", 1)
lib = get_shared_libraries(namespace, version)
if lib:
libs = lib.lower().split(",")
else:
libs = []
for lib in libs:
deps.add((namespace, version, lib))
return deps
def get_dependencies(filename):
deps = []
try:
data = subprocess.check_output(
["objdump", "-p", filename], stderr=subprocess.STDOUT
)
except subprocess.CalledProcessError:
# can happen with wrong arch binaries
return []
data = data.decode("utf-8")
for line in data.splitlines():
line = line.strip()
if line.startswith("DLL Name:"):
deps.append(line.split(":", 1)[-1].strip().lower())
return deps
def find_lib(root, name):
system_search_path = os.path.join("C:", os.sep, "Windows", "System32")
if get_lib_path(root, name):
return True
elif os.path.exists(os.path.join(system_search_path, name)):
return True
elif name in ["gdiplus.dll"]:
return True
elif name.startswith("msvcr"):
return True
return False
def get_lib_path(root, name):
search_path = os.path.join(root, "bin")
if os.path.exists(os.path.join(search_path, name)):
return os.path.join(search_path, name)
def get_things_to_delete(root):
extensions = [".exe", ".pyd", ".dll"]
all_libs = set()
needed = set()
for base, dirs, files in os.walk(root):
for f in files:
lib = f.lower()
path = os.path.join(base, f)
ext_lower = os.path.splitext(f)[-1].lower()
if ext_lower in extensions:
if ext_lower == ".exe":
# we use .exe as dependency root
needed.add(lib)
all_libs.add(f.lower())
for lib in get_dependencies(path):
all_libs.add(lib)
needed.add(lib)
if not find_lib(root, lib):
print("MISSING:", path, lib)
for namespace, version, lib in get_required_by_typelibs():
all_libs.add(lib)
needed.add(lib)
if not find_lib(root, lib):
print("MISSING:", namespace, version, lib)
to_delete = []
for not_depended_on in all_libs - needed:
path = get_lib_path(root, not_depended_on)
if path:
to_delete.append(path)
return to_delete
def main(argv):
libs = get_things_to_delete(sys.prefix)
if "--delete" in argv[1:]:
while libs:
for lib in libs:
print("DELETE:", lib)
os.unlink(lib)
libs = get_things_to_delete(sys.prefix)
if __name__ == "__main__":
main(sys.argv)
|
network | qa_socket_pdu | #!/usr/bin/env python
#
# Copyright 2013 Free Software Foundation, Inc.
#
# This file is part of GNU Radio
#
# SPDX-License-Identifier: GPL-3.0-or-later
#
#
import random
import time
import pmt
from gnuradio import blocks, gr, gr_unittest, network, pdu
class qa_socket_pdu(gr_unittest.TestCase):
def setUp(self):
random.seed(0)
self.tb = gr.top_block()
def tearDown(self):
self.tb = None
def test_001(self):
# Test that blocks can be created and destroyed without hanging
port = str(random.Random().randint(0, 30000) + 10000)
self.pdu_send = network.socket_pdu("UDP_CLIENT", "localhost", port)
self.pdu_recv = network.socket_pdu("UDP_SERVER", "localhost", port)
self.pdu_send = None
self.pdu_recv = None
def test_002(self):
# Send a PDU through a pair of UDP sockets
port = str(random.Random().randint(0, 30000) + 10000)
srcdata = (0x64, 0x6F, 0x67, 0x65)
data = pmt.init_u8vector(srcdata.__len__(), srcdata)
pdu_msg = pmt.cons(pmt.PMT_NIL, data)
self.pdu_source = blocks.message_strobe(pdu_msg, 500)
self.pdu_recv = network.socket_pdu("UDP_SERVER", "localhost", port)
self.pdu_send = network.socket_pdu("UDP_CLIENT", "localhost", port)
self.dbg = blocks.message_debug()
self.tb.msg_connect(self.pdu_source, "strobe", self.pdu_send, "pdus")
self.tb.msg_connect(self.pdu_recv, "pdus", self.dbg, "store")
self.tb.start()
time.sleep(1)
self.tb.stop()
self.tb.wait()
self.pdu_send = None
self.pdu_recv = None
received = self.dbg.get_message(0)
received_data = pmt.cdr(received)
msg_data = []
for i in range(4):
msg_data.append(pmt.u8vector_ref(received_data, i))
self.assertEqual(srcdata, tuple(msg_data))
def test_003(self):
# Test that block stops when interacting with streaming interface
port = str(random.Random().randint(0, 30000) + 10000)
srcdata = (
0x73,
0x75,
0x63,
0x68,
0x74,
0x65,
0x73,
0x74,
0x76,
0x65,
0x72,
0x79,
0x70,
0x61,
0x73,
0x73,
)
tag_dict = {"offset": 0}
tag_dict["key"] = pmt.intern("len")
tag_dict["value"] = pmt.from_long(8)
tag1 = gr.python_to_tag(tag_dict)
tag_dict["offset"] = 8
tag2 = gr.python_to_tag(tag_dict)
tags = [tag1, tag2]
src = blocks.vector_source_b(srcdata, False, 1, tags)
ts_to_pdu = pdu.tagged_stream_to_pdu(gr.types.byte_t, "len")
pdu_send = network.socket_pdu("UDP_CLIENT", "localhost", "4141")
# pdu_recv = network.socket_pdu("UDP_SERVER", "localhost", port)
pdu_to_ts = pdu.pdu_to_tagged_stream(gr.types.byte_t, "len")
head = blocks.head(gr.sizeof_char, 10)
sink = blocks.vector_sink_b(1)
self.tb.connect(src, ts_to_pdu)
self.tb.msg_connect(ts_to_pdu, "pdus", pdu_send, "pdus")
# a UDP socket connects pdu_send to pdu_recv
# TODO: test that the recv socket can be destroyed from downstream
# that signals DONE. Also that we get the PDUs we sent
# self.tb.msg_connect(pdu_recv, "pdus", pdu_to_ts, "pdus")
# self.tb.connect(pdu_to_ts, head, sink)
self.tb.run()
def test_004(self):
# Test that the TCP server can stream PDUs <= the MTU size.
port = str(random.Random().randint(0, 30000) + 10000)
mtu = 10000
srcdata = tuple(x % 256 for x in range(mtu))
data = pmt.init_u8vector(srcdata.__len__(), srcdata)
pdu_msg = pmt.cons(pmt.PMT_NIL, data)
self.pdu_source = blocks.message_strobe(pdu_msg, 500)
self.pdu_send = network.socket_pdu("TCP_SERVER", "localhost", port, mtu)
self.pdu_recv = network.socket_pdu("TCP_CLIENT", "localhost", port, mtu)
self.pdu_sink = blocks.message_debug()
self.tb.msg_connect(self.pdu_source, "strobe", self.pdu_send, "pdus")
self.tb.msg_connect(self.pdu_recv, "pdus", self.pdu_sink, "store")
self.tb.start()
time.sleep(1)
self.tb.stop()
self.tb.wait()
received = self.pdu_sink.get_message(0)
received_data = pmt.cdr(received)
msg_data = []
for i in range(mtu):
msg_data.append(pmt.u8vector_ref(received_data, i))
self.assertEqual(srcdata, tuple(msg_data))
if __name__ == "__main__":
gr_unittest.run(qa_socket_pdu)
|
storage | storage | """
Storing inventory items
"""
import collections
InventoryItem = collections.namedtuple(
"InventoryItem", "type stream payload expires tag"
)
class Storage(object): # pylint: disable=too-few-public-methods
"""Base class for storing inventory
(extendable for other items to store)"""
pass
class InventoryStorage(Storage, collections.MutableMapping):
"""Module used for inventory storage"""
def __init__(self): # pylint: disable=super-init-not-called
self.numberOfInventoryLookupsPerformed = 0
def __contains__(self, _):
raise NotImplementedError
def __getitem__(self, _):
raise NotImplementedError
def __setitem__(self, _, value):
raise NotImplementedError
def __delitem__(self, _):
raise NotImplementedError
def __iter__(self):
raise NotImplementedError
def __len__(self):
raise NotImplementedError
def by_type_and_tag(self, objectType, tag):
"""Return objects filtered by object type and tag"""
raise NotImplementedError
def unexpired_hashes_by_stream(self, stream):
"""Return unexpired inventory vectors filtered by stream"""
raise NotImplementedError
def flush(self):
"""Flush cache"""
raise NotImplementedError
def clean(self):
"""Free memory / perform garbage collection"""
raise NotImplementedError
class MailboxStorage(Storage, collections.MutableMapping):
"""Method for storing mails"""
def __delitem__(self, key):
raise NotImplementedError
def __getitem__(self, key):
raise NotImplementedError
def __iter__(self):
raise NotImplementedError
def __len__(self):
raise NotImplementedError
def __setitem__(self, key, value):
raise NotImplementedError
|
brainz | mb | # Copyright 2016 Christoph Reiter
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
try:
import musicbrainzngs
except ImportError:
from quodlibet import plugins
raise plugins.MissingModulePluginException("musicbrainzngs")
from quodlibet import app, const, util
VARIOUS_ARTISTS_ARTISTID = "89ad4ac3-39f7-470e-963a-56509c546377"
def is_mbid(id_):
return len(id_) == 36
def search_releases(query):
"""Returns a list of ReleaseResult or raises MusicBrainzError"""
musicbrainzngs.set_useragent(app.name, const.VERSION)
return [Release(r) for r in musicbrainzngs.search_releases(query)["release-list"]]
def _get_release(release_id):
"""Returns a release containing all recordings and artists or raises
MusicBrainzError
"""
assert is_mbid(release_id)
return musicbrainzngs.get_release_by_id(
release_id, includes=["recordings", "artists", "artist-credits", "labels"]
)["release"]
class Artist:
def __init__(self, name, sort_name, id_):
self.name = name
self.sort_name = sort_name
self.id = id_ # MusicBrainz Artist ID / Album Artist ID
@property
def is_various(self):
return self.id == VARIOUS_ARTISTS_ARTISTID
@classmethod
def from_credit(cls, mbcredit):
artists = []
for credit in mbcredit:
try:
artist = credit["artist"]
except TypeError:
# join strings
pass
else:
artists.append(
Artist(artist["name"], artist["sort-name"], artist["id"])
)
return artists
class ReleaseTrack:
"""Part of a Release, combines per track and per medium data"""
def __init__(self, mbtrack, discnumber, track_count, disctitle):
self._mbtrack = mbtrack
self.discnumber = discnumber
self.track_count = track_count
self.disctitle = disctitle
@property
def id(self):
"""MusicBrainz release track ID"""
return self._mbtrack["id"]
@property
def artists(self):
return Artist.from_credit(self._mbtrack["artist-credit"])
@property
def title(self):
return self._mbtrack["recording"]["title"]
@property
def tracknumber(self):
return self._mbtrack["position"]
class Release:
def __init__(self, mbrelease):
self._mbrelease = mbrelease
@property
def labelid(self):
label_list = self._mbrelease.get("label-info-list", [])
if not label_list:
return ""
return label_list[0].get("catalog-number", "")
@property
def id(self):
"""MusicBrainz release ID"""
return self._mbrelease["id"]
@property
def date(self):
return self._mbrelease.get("date", "")
@property
def medium_format(self):
formats = []
for medium in self._mbrelease["medium-list"]:
format_ = medium.get("format", "")
if format_:
formats.append(format_)
formats = util.list_unique(formats)
return "/".join(formats)
@property
def country(self):
return self._mbrelease.get("country", "")
@property
def disc_count(self):
return self._mbrelease["medium-count"]
@property
def track_count(self):
"""Number of tracks for all included mediums"""
track_count = 0
for medium in self._mbrelease["medium-list"]:
if "track-count" in medium:
track_count += medium["track-count"]
if "pregap" in medium:
track_count += 1
return track_count
@property
def tracks(self):
tracks = []
for medium in self._mbrelease["medium-list"]:
disc = medium["position"]
title = medium.get("title", "")
track_count = medium["track-count"]
if "pregap" in medium:
track_count += 1
tracks.append(ReleaseTrack(medium["pregap"], disc, track_count, title))
for track in medium["track-list"]:
tracks.append(ReleaseTrack(track, disc, track_count, title))
return tracks
@property
def title(self):
return self._mbrelease["title"]
@property
def is_single_artist(self):
"""If all tracks have the same artists as the release"""
ids = [a.id for a in self.artists]
for track in self.tracks:
track_ids = [a.id for a in track.artists]
if ids != track_ids:
return False
return True
@property
def is_various_artists(self):
artists = self.artists
return len(artists) == 1 and artists[0].is_various
@property
def artists(self):
return Artist.from_credit(self._mbrelease["artist-credit"])
def fetch_full(self):
"""Returns a new Release instance containing more info
or raises MusicBrainzError
This method is blocking...
"""
return Release(_get_release(self.id))
|
utils | compat | """
raven.utils.compat
~~~~~~~~~~~~~~~~~~
:copyright: (c) 2010-2016 by the Sentry Team, see AUTHORS for more details.
:license: BSD, see LICENSE for more details.
Utilities for writing code that runs on Python 2 and 3
"""
# flake8: noqa
# Copyright (c) 2010-2013 Benjamin Peterson
#
# Permission is hereby granted, free of charge, to any person obtaining a copy of
# this software and associated documentation files (the "Software"), to deal in
# the Software without restriction, including without limitation the rights to
# use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
# the Software, and to permit persons to whom the Software is furnished to do so,
# subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
# FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
# COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
# IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
# CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
from __future__ import absolute_import
import operator
import sys
import types
PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3
if PY3:
string_types = (str,)
integer_types = (int,)
class_types = (type,)
text_type = str
binary_type = bytes
else:
string_types = (basestring,)
integer_types = (int, long)
class_types = (type, types.ClassType)
text_type = unicode
binary_type = str
try:
advance_iterator = next
except NameError:
def advance_iterator(it):
return it.next()
next = advance_iterator
try:
callable = callable
except NameError:
def callable(obj):
return any("__call__" in klass.__dict__ for klass in type(obj).__mro__)
if PY3:
Iterator = object
else:
class Iterator:
def next(self):
return type(self).__next__(self)
if PY3:
def iterkeys(d, **kw):
return iter(d.keys(**kw))
def itervalues(d, **kw):
return iter(d.values(**kw))
def iteritems(d, **kw):
return iter(d.items(**kw))
def iterlists(d, **kw):
return iter(d.lists(**kw))
else:
def iterkeys(d, **kw):
return d.iterkeys(**kw)
def itervalues(d, **kw):
return d.itervalues(**kw)
def iteritems(d, **kw):
return d.iteritems(**kw)
def iterlists(d, **kw):
return d.iterlists(**kw)
if PY3:
def b(s):
return s.encode("latin-1")
def u(s):
return s
import io
StringIO = io.StringIO
BytesIO = io.BytesIO
else:
def b(s):
return s
# Workaround for standalone backslash
def u(s):
return unicode(s.replace(r"\\", r"\\\\"), "unicode_escape")
import StringIO
StringIO = BytesIO = StringIO.StringIO
if PY3:
exec_ = getattr(__import__("builtins"), "exec")
def reraise(tp, value, tb=None):
try:
if value is None:
value = tp()
if value.__traceback__ is not tb:
raise value.with_traceback(tb)
raise value
finally:
value = None
tb = None
else:
def exec_(_code_, _globs_=None, _locs_=None):
"""Execute code in a namespace."""
if _globs_ is None:
frame = sys._getframe(1)
_globs_ = frame.f_globals
if _locs_ is None:
_locs_ = frame.f_locals
del frame
elif _locs_ is None:
_locs_ = _globs_
exec("""exec _code_ in _globs_, _locs_""")
exec_(
"""def reraise(tp, value, tb=None):
try:
raise tp, value, tb
finally:
tb = None
"""
)
if sys.version_info[:2] == (3, 2):
exec_(
"""def raise_from(value, from_value):
try:
if from_value is None:
raise value
raise value from from_value
finally:
value = None
"""
)
elif sys.version_info[:2] > (3, 2):
exec_(
"""def raise_from(value, from_value):
try:
raise value from from_value
finally:
value = None
"""
)
else:
def raise_from(value, from_value):
raise value
if PY3:
import urllib.request as urllib2
from http import client as httplib
from queue import Queue
from urllib import parse as urlparse
from urllib.error import HTTPError
from urllib.parse import quote as urllib_quote
else:
from urllib import quote as urllib_quote
import httplib
import urllib2
import urlparse
from Queue import Queue
from urllib2 import HTTPError
def get_code(func):
rv = getattr(func, "__code__", getattr(func, "func_code", None))
if rv is None:
raise TypeError("Could not get code from %r" % type(func).__name__)
return rv
def check_threads():
try:
from uwsgi import opt
except ImportError:
return
# When `threads` is passed in as a uwsgi option,
# `enable-threads` is implied on.
if "threads" in opt:
return
if str(opt.get("enable-threads", "0")).lower() in ("false", "off", "no", "0"):
from warnings import warn
warn(
Warning(
"We detected the use of uwsgi with disabled threads. "
"This will cause issues with the transport you are "
"trying to use. Please enable threading for uwsgi. "
'(Enable the "enable-threads" flag).'
)
)
|
Assembly | TestAssemblyWorkbench | # SPDX-License-Identifier: LGPL-2.1-or-later
# /****************************************************************************
# *
# Copyright (c) 2023 Ondsel <development@ondsel.com> *
# *
# This file is part of FreeCAD. *
# *
# FreeCAD is free software: you can redistribute it and/or modify it *
# under the terms of the GNU Lesser General Public License as *
# published by the Free Software Foundation, either version 2.1 of the *
# License, or (at your option) any later version. *
# *
# FreeCAD is distributed in the hope that it will be useful, but *
# WITHOUT ANY WARRANTY; without even the implied warranty of *
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU *
# Lesser General Public License for more details. *
# *
# You should have received a copy of the GNU Lesser General Public *
# License along with FreeCAD. If not, see *
# <https://www.gnu.org/licenses/>. *
# *
# ***************************************************************************/
import TestApp
from AssemblyTests.TestCore import TestCore
# dummy usage to get flake8 and lgtm quiet
False if TestCore.__name__ else True
False if TestApp.__name__ else True
|
extensions | ubuntu_unity | # -*- coding: utf-8 -*-
# Ubuntu Unity Launcher Integration
# Thomas Perl <thp@gpodder.org>; 2012-02-06
import logging
import gpodder
import gi # isort:skip
gi.require_version("Unity", "7.0") # isort:skip
from gi.repository import GLib, Unity # isort:skip
_ = gpodder.gettext
logger = logging.getLogger(__name__)
__title__ = _("Ubuntu Unity Integration")
__description__ = _("Show download progress in the Unity Launcher icon.")
__authors__ = "Thomas Perl <thp@gpodder.org>"
__category__ = "desktop-integration"
__only_for__ = "unity"
__mandatory_in__ = "unity"
__disable_in__ = "win32"
class LauncherEntry:
FILENAME = "gpodder.desktop"
def __init__(self):
self.launcher = Unity.LauncherEntry.get_for_desktop_id(self.FILENAME)
def set_count(self, count):
self.launcher.set_property("count", count)
self.launcher.set_property("count_visible", count > 0)
def set_progress(self, progress):
self.launcher.set_property("progress", progress)
self.launcher.set_property("progress_visible", 0.0 <= progress < 1.0)
class gPodderExtension:
FILENAME = "gpodder.desktop"
def __init__(self, container):
self.container = container
self.launcher_entry = None
def on_load(self):
logger.info("Starting Ubuntu Unity Integration.")
self.launcher_entry = LauncherEntry()
def on_unload(self):
self.launcher_entry = None
def on_download_progress(self, progress):
GLib.idle_add(self.launcher_entry.set_progress, float(progress))
|
neubot | bytegen_speedtest | # neubot/bytegen_speedtest.py
#
# Copyright (c) 2012 Simone Basso <bassosimone@gmail.com>,
# NEXA Center for Internet & Society at Politecnico di Torino
#
# This file is part of Neubot <http://www.neubot.org/>.
#
# Neubot is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Neubot is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Neubot. If not, see <http://www.gnu.org/licenses/>.
#
""" Generates bytes for the speedtest test """
import getopt
import sys
if __name__ == "__main__":
sys.path.insert(0, ".")
from neubot import utils
from neubot.utils_random import RANDOMBLOCKS
PIECE_LEN = 262144
class BytegenSpeedtest(object):
"""Bytes generator for speedtest"""
def __init__(self, seconds, piece_len=PIECE_LEN):
"""Initializer"""
self.seconds = seconds
self.ticks = utils.ticks()
self.closed = False
self.piece_len = piece_len
def read(self, count=sys.maxint):
"""Read count bytes"""
if self.closed:
return ""
if count < self.piece_len:
raise RuntimeError("Invalid count")
diff = utils.ticks() - self.ticks
if diff < self.seconds:
data = RANDOMBLOCKS.get_block()[: self.piece_len]
length = "%x\r\n" % self.piece_len
vector = [length, data, "\r\n"]
else:
vector = ["0\r\n", "\r\n"]
self.closed = True
return "".join(vector)
def close(self):
"""Close"""
self.closed = True
def main(args):
"""Main() function"""
try:
options, arguments = getopt.getopt(args[1:], "t:")
except getopt.error:
sys.exit("usage: neubot bytegen_speedtest [-t seconds]")
if arguments:
sys.exit("usage: neubot bytegen_speedtest [-t seconds]")
seconds = 5.0
for name, value in options:
if name == "-t":
seconds = float(value)
bytegen = BytegenSpeedtest(seconds)
while True:
data = bytegen.read()
if not data:
break
sys.stdout.write(data)
if __name__ == "__main__":
main(sys.argv)
|
sandbox | test_subsampler | import matplotlib.pyplot as plt
import numpy as np
from friture import generated_filters
from friture.audiobackend import SAMPLING_RATE
from friture.delay_estimator import subsampler, subsampler_filtic
Ns = int(1e4)
# y = np.random.rand(Ns)
f = 2e1
t = np.linspace(0, float(Ns) / SAMPLING_RATE, Ns)
y = np.cos(2.0 * np.pi * f * t)
Ndec = 2
subsampled_sampling_rate = SAMPLING_RATE / 2 ** (Ndec)
[bdec, adec] = generated_filters.PARAMS["dec"]
zfs0 = subsampler_filtic(Ndec, bdec, adec)
Nb = 10
l = int(Ns / Nb)
Nb0 = Nb / 2
print("Nb0 =", Nb0, "Nb1 =", Nb - Nb0)
print("subsample first parts")
for i in range(Nb0):
ydec, zfs0 = subsampler(Ndec, bdec, adec, y[i * l : (i + 1) * l], zfs0)
if i == 0:
y_dec = ydec
else:
y_dec = np.append(y_dec, ydec)
print(
"push an empty array to the subsampler",
y[i * l : i * l].shape,
y[i * l : i * l].size,
)
ydec, zfs0 = subsampler(Ndec, bdec, adec, y[i * l : i * l], zfs0)
y_dec = np.append(y_dec, ydec)
print("subsample last parts")
for i in range(Nb0, Nb):
ydec, zfs0 = subsampler(Ndec, bdec, adec, y[i * l : (i + 1) * l], zfs0)
y_dec = np.append(y_dec, ydec)
print("plot")
plt.figure()
plt.subplot(2, 1, 1)
plt.plot(y)
plt.subplot(2, 1, 2)
plt.plot(y_dec)
plt.show()
|
PyObjCTest | test_nsbitmapimagerep | import array
import sys
import objc
from AppKit import *
from objc import NO, YES
from PyObjCTools.TestSupport import *
try:
unicode
except NameError:
unicode = str
try:
long
except NameError:
long = int
class TestNSBitmapImageRep(TestCase):
def testInstantiation(self):
# widthxheight RGB 24bpp image
width = 256
height = 256
dataPlanes = (None, None, None, None, None)
dataPlanes = None
i1 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
dataPlanes, width, height, 8, 3, NO, NO, NSDeviceRGBColorSpace, 0, 0
)
self.assertTrue(i1)
i2 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
None, width, height, 8, 3, NO, NO, NSDeviceRGBColorSpace, 0, 0
)
self.assertTrue(i2)
def testPixelFormat(self):
width = 16
height = 16
i1 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bitmapFormat_bytesPerRow_bitsPerPixel_(
None,
width,
height,
8,
3,
NO,
NO,
NSDeviceRGBColorSpace,
NSAlphaFirstBitmapFormat,
0,
0,
)
self.assertIsInstance(i1, NSBitmapImageRep)
singlePlane = objc.allocateBuffer(width * height * 4)
for i in range(0, width * height):
si = i * 4
singlePlane[si] = 1
singlePlane[si + 1] = 2
singlePlane[si + 2] = 3
singlePlane[si + 3] = 4
dataPlanes = (singlePlane, None, None, None, None)
# test non-planar, premade buffer
i2 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bitmapFormat_bytesPerRow_bitsPerPixel_(
dataPlanes,
width,
height,
8,
3,
NO,
NO,
NSDeviceRGBColorSpace,
NSAlphaFirstBitmapFormat,
0,
0,
)
self.assertIsInstance(i2, NSBitmapImageRep)
bitmapData = i2.bitmapData()
self.assertEqual(len(bitmapData), width * height * 4)
def testImageData(self):
width = 256
height = 256
rPlane = array.array("B")
rPlane.fromlist([y % 256 for y in range(0, height) for x in range(0, width)])
if sys.version_info[0] == 3:
buffer = memoryview
else:
from __builtin__ import buffer
rPlane = buffer(rPlane)
gPlane = array.array("B")
gPlane.fromlist(
[y % 256 for y in range(0, height) for x in range(width, 0, -1)]
)
gPlane = buffer(gPlane)
bPlane = array.array("B")
bPlane.fromlist([x % 256 for y in range(0, height) for x in range(0, width)])
bPlane = buffer(bPlane)
dataPlanes = (rPlane, gPlane, bPlane, None, None)
# test planar, pre-made buffer
i1 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
dataPlanes, width, height, 8, 3, NO, YES, NSDeviceRGBColorSpace, 0, 0
)
self.assertTrue(i1)
singlePlane = objc.allocateBuffer(width * height * 3)
for i in range(0, width * height):
si = i * 3
if sys.version_info[0] == 2:
singlePlane[si] = rPlane[i]
singlePlane[si + 1] = gPlane[i]
singlePlane[si + 2] = bPlane[i]
else:
def as_byte(v):
if isinstance(v, int):
return v
else:
return ord(v)
singlePlane[si] = as_byte(rPlane[i])
singlePlane[si + 1] = as_byte(gPlane[i])
singlePlane[si + 2] = as_byte(bPlane[i])
dataPlanes = (singlePlane, None, None, None, None)
# test non-planar, premade buffer
i2 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
dataPlanes, width, height, 8, 3, NO, NO, NSDeviceRGBColorSpace, 0, 0
)
# test grey scale
greyPlane = array.array("B")
greyPlane.fromlist([x % 256 for x in range(0, height) for x in range(0, width)])
greyPlanes = (greyPlane, None, None, None, None)
greyImage = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
greyPlanes,
width,
height,
8,
1,
NO,
YES,
NSCalibratedWhiteColorSpace,
width,
8,
)
# test planar, NSBIR allocated buffer
i3 = NSBitmapImageRep.alloc().initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
None, width, height, 8, 3, NO, YES, NSDeviceRGBColorSpace, 0, 0
)
r, g, b, a, o = i3.getBitmapDataPlanes_()
self.assertTrue(r)
self.assertTrue(g)
self.assertTrue(b)
self.assertTrue(not a)
self.assertTrue(not o)
self.assertEqual(len(r), len(rPlane))
self.assertEqual(len(g), len(gPlane))
self.assertEqual(len(b), len(bPlane))
r[0 : len(r)] = rPlane[0 : len(rPlane)]
g[0 : len(g)] = gPlane[0 : len(gPlane)]
b[0 : len(b)] = bPlane[0 : len(bPlane)]
bitmapData = i2.bitmapData()
self.assertEqual(len(bitmapData), len(singlePlane))
try:
memoryview
except NameError:
self.assertEqual(bitmapData, singlePlane)
else:
self.assertEqual(bitmapData.tobytes(), singlePlane)
a = array.array("L", [255] * 4)
self.assertArgIsOut(NSBitmapImageRep.getPixel_atX_y_, 0)
d = i2.getPixel_atX_y_(a, 1, 1)
self.assertIs(a, d)
class TestBadCreation(TestCase):
# Redirect stderr to /dev/null for the duration of this test,
# NSBitmapImageRep will write an error message to stderr.
def setUp(self):
import os
self.duppedStderr = os.dup(2)
fp = os.open("/dev/null", os.O_RDWR)
os.dup2(fp, 2)
os.close(fp)
def tearDown(self):
import os
os.dup2(self.duppedStderr, 2)
def test_AllocInit(self):
y = NSBitmapImageRep.alloc()
try:
self.assertRaises(ValueError, y.init)
finally:
width = 256
height = 256
dataPlanes = (None, None, None, None, None)
y = y.initWithBitmapDataPlanes_pixelsWide_pixelsHigh_bitsPerSample_samplesPerPixel_hasAlpha_isPlanar_colorSpaceName_bytesPerRow_bitsPerPixel_(
dataPlanes, width, height, 8, 3, NO, NO, NSDeviceRGBColorSpace, 0, 0
)
def testConstants(self):
self.assertEqual(NSTIFFCompressionNone, 1)
self.assertEqual(NSTIFFCompressionCCITTFAX3, 3)
self.assertEqual(NSTIFFCompressionCCITTFAX4, 4)
self.assertEqual(NSTIFFCompressionLZW, 5)
self.assertEqual(NSTIFFCompressionJPEG, 6)
self.assertEqual(NSTIFFCompressionNEXT, 32766)
self.assertEqual(NSTIFFCompressionPackBits, 32773)
self.assertEqual(NSTIFFCompressionOldJPEG, 32865)
self.assertEqual(NSTIFFFileType, 0)
self.assertEqual(NSBMPFileType, 1)
self.assertEqual(NSGIFFileType, 2)
self.assertEqual(NSJPEGFileType, 3)
self.assertEqual(NSPNGFileType, 4)
self.assertEqual(NSJPEG2000FileType, 5)
self.assertEqual(NSImageRepLoadStatusUnknownType, -1)
self.assertEqual(NSImageRepLoadStatusReadingHeader, -2)
self.assertEqual(NSImageRepLoadStatusWillNeedAllData, -3)
self.assertEqual(NSImageRepLoadStatusInvalidData, -4)
self.assertEqual(NSImageRepLoadStatusUnexpectedEOF, -5)
self.assertEqual(NSImageRepLoadStatusCompleted, -6)
self.assertEqual(NSAlphaFirstBitmapFormat, 1 << 0)
self.assertEqual(NSAlphaNonpremultipliedBitmapFormat, 1 << 1)
self.assertEqual(NSFloatingPointSamplesBitmapFormat, 1 << 2)
self.assertIsInstance(NSImageCompressionMethod, unicode)
self.assertIsInstance(NSImageCompressionFactor, unicode)
self.assertIsInstance(NSImageDitherTransparency, unicode)
self.assertIsInstance(NSImageRGBColorTable, unicode)
self.assertIsInstance(NSImageInterlaced, unicode)
self.assertIsInstance(NSImageColorSyncProfileData, unicode)
self.assertIsInstance(NSImageFrameCount, unicode)
self.assertIsInstance(NSImageCurrentFrame, unicode)
self.assertIsInstance(NSImageCurrentFrameDuration, unicode)
self.assertIsInstance(NSImageLoopCount, unicode)
self.assertIsInstance(NSImageGamma, unicode)
self.assertIsInstance(NSImageProgressive, unicode)
self.assertIsInstance(NSImageEXIFData, unicode)
self.assertIsInstance(NSImageFallbackBackgroundColor, unicode)
def testTiffCompression(self):
lst, nr = NSBitmapImageRep.getTIFFCompressionTypes_count_(None, None)
self.assertIsInstance(lst, tuple)
self.assertIsInstance(nr, (int, long))
self.assertEqual(len(lst), nr)
self.assertNotEqual(len(lst), 0)
self.assertIsInstance(lst[0], (int, long))
def testMethods(self):
self.assertResultIsBOOL(NSBitmapImageRep.isPlanar)
self.assertResultIsBOOL(NSBitmapImageRep.canBeCompressedUsing_)
self.assertArgIsBOOL(NSBitmapImageRep.incrementalLoadFromData_complete_, 1)
self.assertArgIsOut(NSBitmapImageRep.getCompression_factor_, 0)
self.assertArgIsOut(NSBitmapImageRep.getCompression_factor_, 1)
if __name__ == "__main__":
main()
|
imports | troubleshoot | """ import books from another app """
from bookwyrm import models
from bookwyrm.importers import Importer
from bookwyrm.settings import PAGE_LENGTH
from django.contrib.auth.decorators import login_required
from django.core.exceptions import PermissionDenied
from django.core.paginator import Paginator
from django.shortcuts import get_object_or_404, redirect
from django.template.response import TemplateResponse
from django.urls import reverse
from django.utils.decorators import method_decorator
from django.views import View
# pylint: disable= no-self-use
@method_decorator(login_required, name="dispatch")
class ImportTroubleshoot(View):
"""problems items in an existing import"""
def get(self, request, job_id):
"""status of an import job"""
job = get_object_or_404(models.ImportJob, id=job_id)
if job.user != request.user:
raise PermissionDenied()
items = job.items.order_by("index").filter(
fail_reason__isnull=False, book_guess__isnull=True
)
paginated = Paginator(items, PAGE_LENGTH)
page = paginated.get_page(request.GET.get("page"))
data = {
"job": job,
"items": page,
"page_range": paginated.get_elided_page_range(
page.number, on_each_side=2, on_ends=1
),
"complete": True,
"page_path": reverse("import-troubleshoot", args=[job.id]),
}
return TemplateResponse(request, "import/troubleshoot.html", data)
def post(self, request, job_id):
"""retry lines from an import"""
job = get_object_or_404(models.ImportJob, id=job_id)
items = job.items.filter(fail_reason__isnull=False)
importer = Importer()
job = importer.create_retry_job(
request.user,
job,
items,
)
job.start_job()
return redirect(f"/import/{job.id}")
|
puddlestuff | about | # -*- coding: utf-8 -*-
from importlib import import_module
from platform import python_version
import mutagen
from PyQt5.QtCore import PYQT_VERSION_STR, QT_VERSION_STR, Qt
from PyQt5.QtGui import QPixmap
from PyQt5.QtWidgets import (
QApplication,
QDialog,
QHBoxLayout,
QLabel,
QScrollArea,
QTabWidget,
QVBoxLayout,
QWidget,
)
from . import changeset, version_string
from .puddleobjects import OKCancel
from .translations import translate
desc = translate(
"About",
"""puddletag is an audio tag editor for GNU/Linux similar to the Windows program Mp3tag.
<br /><br />Features include: Batch editing of tags, renaming files using tags, retrieving tags from filenames, using Actions to automate repetitive tasks, importing your music library and loads of other awesome stuff. <br /><br />
Supported formats: id3v1, id3v2 (.mp3), AAC (.mp4, .m4a), VorbisComments (.ogg, .flac) and APEv2 (.ape) <br />< br />
Visit the puddletag website (<a href="http://puddletag.sourceforge.net">http://puddletag.sourceforge.net</a>) for help and updates.<br /><br />
© 2008-2012 concentricpuddle (concentricpuddle@gmail.com) <br />
Licensed under GPLv3 (<a href="www.gnu.org/licenses/gpl-3.0.html">www.gnu.org/licenses/gpl-3.0.html</a>).
""",
)
thanks = translate(
"About",
"""<b>Evan Devetzis</b> for his many, many awesome ideas and putting up with more bugs than humanly possible.<br /><br />
First off, a big thanks to **Evan Devetzis** for working tirelessly in helping me make puddletag better by contributing many, many awesome ideas and for being a great bug hunter.
Thanks to <b>Raphaël Rochet</b>, <b>Fabian Bakkum</b>, <b>Alan Gomes</b> and others for contributing translations.
To the writers of the libraries puddletag depends on (without which I'll probably still be writing an id3 reader).<br /><br />
<b>Paul McGuire</b> for PyParsing.<br />
<b>Michael Urman</b> and <b>Joe Wreschnig</b> for Mutagen (It. Is. Awesome).<br />
<b>Phil Thomson</b> and everyone responsible for PyQt4.<br />
<b>Michael Foord</b> and <b>Nicola Larosa</b> for ConfigObj (seriously, they should replace ConfigParser with this).<br />
The <b>Oxygen team</b> for the Oxygen icons.
""",
)
def versions():
def get_module_version(module_name):
try:
from importlib.metadata import version
return version(module_name)
except ModuleNotFoundError:
pass
try:
module = import_module(module_name)
return getattr(module, "__version__", translate("About", "unknown version"))
except ModuleNotFoundError:
return translate("About", "not installed")
return {
"Python": python_version(),
"PyQt": PYQT_VERSION_STR,
"Qt": QT_VERSION_STR,
"Mutagen": mutagen.version_string,
"PyParsing": get_module_version("pyparsing"),
"ConfigObj": get_module_version("configobj"),
"Unidecode": get_module_version("unidecode"),
"lxml": get_module_version("lxml"),
"pyacoustid": get_module_version("pyacoustid"),
"audioread": get_module_version("audioread"),
"Levenshtein": get_module_version("Levenshtein"),
"Chromaprint": get_module_version("chromaprint"),
}
class ScrollLabel(QWidget):
def __init__(self, text, alignment=Qt.AlignmentFlag.AlignCenter, parent=None):
QWidget.__init__(self, parent)
vbox = QVBoxLayout()
self.setLayout(vbox)
label = QLabel(text)
label.setTextFormat(Qt.TextFormat.RichText)
label.setAlignment(alignment)
label.setWordWrap(True)
label.setTextInteractionFlags(Qt.TextInteractionFlag.TextBrowserInteraction)
sa = QScrollArea()
sa.setWidget(label)
sa.setWidgetResizable(True)
vbox.addWidget(sa)
self.label = label
class AboutPuddletag(QDialog):
def __init__(self, parent=None):
QDialog.__init__(self, parent)
self.setWindowTitle(translate("About", "About puddletag"))
icon = QLabel()
icon.setPixmap(QPixmap(":/appicon.svg").scaled(48, 48))
lib_versions = "<br />".join(
["%s: %s" % (lib, version) for (lib, version) in versions().items()]
)
if changeset:
version = translate("About", "<h2>puddletag %1</h2>Changeset %2")
version = version.arg(version_string)
version = version.arg(changeset)
else:
version = translate("About", "<h2>puddletag %1</h2>")
version = version.arg(version_string)
label = QLabel(version)
tab = QTabWidget()
tab.addTab(ScrollLabel(desc), translate("About", "&About"))
tab.addTab(
ScrollLabel(thanks, Qt.AlignmentFlag.AlignLeft),
translate("About", "&Thanks"),
)
tab.addTab(
ScrollLabel(lib_versions, Qt.AlignmentFlag.AlignLeft),
translate("About", "&Libraries"),
)
vbox = QVBoxLayout()
version_layout = QHBoxLayout()
version_layout.addWidget(icon)
version_layout.addWidget(label, 1)
vbox.addLayout(version_layout)
vbox.addWidget(tab, 1)
ok = OKCancel()
ok.cancelButton.setVisible(False)
vbox.addLayout(ok)
ok.ok.connect(self.close)
self.setLayout(vbox)
if __name__ == "__main__":
app = QApplication([])
win = AboutPuddletag()
win.show()
app.exec_()
|
engines | mediawiki | # SPDX-License-Identifier: AGPL-3.0-or-later
"""
General mediawiki-engine (Web)
"""
from json import loads
from string import Formatter
from urllib.parse import quote, urlencode
# about
about = {
"website": None,
"wikidata_id": None,
"official_api_documentation": "http://www.mediawiki.org/wiki/API:Search",
"use_official_api": True,
"require_api_key": False,
"results": "JSON",
}
# engine dependent config
categories = ["general"]
paging = True
number_of_results = 1
search_type = "nearmatch" # possible values: title, text, nearmatch
# search-url
base_url = "https://{language}.wikipedia.org/"
search_postfix = (
"w/api.php?action=query"
"&list=search"
"&{query}"
"&format=json"
"&sroffset={offset}"
"&srlimit={limit}"
"&srwhat={searchtype}"
)
# do search-request
def request(query, params):
offset = (params["pageno"] - 1) * number_of_results
string_args = dict(
query=urlencode({"srsearch": query}),
offset=offset,
limit=number_of_results,
searchtype=search_type,
)
format_strings = list(Formatter().parse(base_url))
if params["language"] == "all":
language = "en"
else:
language = params["language"].split("-")[0]
# format_string [('https://', 'language', '', None), ('.wikipedia.org/', None, None, None)]
if any(x[1] == "language" for x in format_strings):
string_args["language"] = language
# write search-language back to params, required in response
params["language"] = language
search_url = base_url + search_postfix
params["url"] = search_url.format(**string_args)
return params
# get response from search-request
def response(resp):
results = []
search_results = loads(resp.text)
# return empty array if there are no results
if not search_results.get("query", {}).get("search"):
return []
# parse results
for result in search_results["query"]["search"]:
if result.get("snippet", "").startswith("#REDIRECT"):
continue
url = (
base_url.format(language=resp.search_params["language"])
+ "wiki/"
+ quote(result["title"].replace(" ", "_").encode())
)
# append result
results.append({"url": url, "title": result["title"], "content": ""})
# return results
return results
|
bukuserver | api | #!/usr/bin/env python
# pylint: disable=wrong-import-order, ungrouped-imports
"""Server module."""
import collections
import typing as T
from unittest import mock
import buku
import flask
from buku import BukuDb
from flask import current_app, jsonify, redirect, request, url_for
from flask.views import MethodView
from flask_api import exceptions, status
try:
from . import forms, response
except ImportError:
from bukuserver import forms, response
STATISTIC_DATA = None
response_ok = lambda: (
jsonify(response.response_template["success"]),
status.HTTP_200_OK,
{"ContentType": "application/json"},
)
response_bad = lambda: (
jsonify(response.response_template["failure"]),
status.HTTP_400_BAD_REQUEST,
{"ContentType": "application/json"},
)
to_response = lambda ok: response_ok() if ok else response_bad()
def entity(bookmark, id=False):
data = {
"id": bookmark.id,
"url": bookmark.url,
"title": bookmark.title,
"tags": bookmark.taglist,
"description": bookmark.desc,
}
if not id:
data.pop("id")
return data
def get_bukudb():
"""get bukudb instance"""
db_file = current_app.config.get("BUKUSERVER_DB_FILE", None)
return BukuDb(dbfile=db_file)
def search_tag(
db: BukuDb, stag: T.Optional[str] = None, limit: T.Optional[int] = None
) -> T.Tuple[T.List[str], T.Dict[str, int]]:
"""search tag.
db:
buku db instance
stag:
search tag
limit:
positive integer limit
Returns
-------
tuple
list of unique tags sorted alphabetically and dictionary of tag and its usage count
Raises
------
ValueError
if limit is not positive
"""
if limit is not None and limit < 1:
raise ValueError("limit must be positive")
tags: T.Set[str] = set()
counter = collections.Counter()
query_list = ["SELECT DISTINCT tags , COUNT(tags) FROM bookmarks"]
if stag:
query_list.append("where tags LIKE :search_tag")
query_list.append("GROUP BY tags")
row: T.Tuple[str, int]
for row in db.cur.execute(" ".join(query_list), {"search_tag": f"%{stag}%"}):
for tag in row[0].strip(buku.DELIM).split(buku.DELIM):
if not tag:
continue
tags.add(tag)
counter[tag] += row[1]
return list(sorted(tags)), dict(counter.most_common(limit))
class ApiTagView(MethodView):
def get(self, tag: T.Optional[str]):
bukudb = get_bukudb()
if tag is None:
return {"tags": search_tag(db=bukudb, limit=5)[0]}
tags = search_tag(db=bukudb, stag=tag)
if tag not in tags[1]:
raise exceptions.NotFound()
return {"name": tag, "usage_count": tags[1][tag]}
def put(self, tag: str):
bukudb = get_bukudb()
try:
new_tags = request.data.get("tags") # type: ignore
if new_tags:
new_tags = new_tags.split(",")
else:
return response_bad()
except AttributeError as e:
raise exceptions.ParseError(detail=str(e))
return to_response(bukudb.replace_tag(tag, new_tags))
class ApiBookmarkView(MethodView):
def get(self, rec_id: T.Union[int, None]):
if rec_id is None:
bukudb = getattr(flask.g, "bukudb", get_bukudb())
all_bookmarks = bukudb.get_rec_all()
result = {
"bookmarks": [
entity(bookmark, id=not request.path.startswith("/api/"))
for bookmark in all_bookmarks
]
}
res = jsonify(result)
else:
bukudb = getattr(flask.g, "bukudb", get_bukudb())
bookmark = bukudb.get_rec_by_id(rec_id)
res = response_bad() if bookmark is None else jsonify(entity(bookmark))
return res
def post(self, rec_id: None = None):
bukudb = getattr(flask.g, "bukudb", get_bukudb())
create_bookmarks_form = forms.ApiBookmarkForm()
url_data = create_bookmarks_form.url.data
result_flag = bukudb.add_rec(
url_data,
create_bookmarks_form.title.data,
create_bookmarks_form.tags.data,
create_bookmarks_form.description.data,
)
return to_response(result_flag)
def put(self, rec_id: int):
bukudb = getattr(flask.g, "bukudb", get_bukudb())
result_flag = bukudb.update_rec(
rec_id,
request.form.get("url"),
request.form.get("title"),
request.form.get("tags"),
request.form.get("description"),
)
return to_response(result_flag)
def delete(self, rec_id: T.Union[int, None]):
if rec_id is None:
bukudb = getattr(flask.g, "bukudb", get_bukudb())
with mock.patch("buku.read_in", return_value="y"):
result_flag = bukudb.cleardb()
else:
bukudb = getattr(flask.g, "bukudb", get_bukudb())
result_flag = bukudb.delete_rec(rec_id)
return to_response(result_flag)
class ApiBookmarkRangeView(MethodView):
def get(self, starting_id: int, ending_id: int):
bukudb = getattr(flask.g, "bukudb", get_bukudb())
max_id = bukudb.get_max_id() or 0
if starting_id > max_id or ending_id > max_id:
return response_bad()
result = {
"bookmarks": {
i: entity(bukudb.get_rec_by_id(i))
for i in range(starting_id, ending_id + 1)
}
}
return jsonify(result)
def put(self, starting_id: int, ending_id: int):
bukudb = getattr(flask.g, "bukudb", get_bukudb())
max_id = bukudb.get_max_id() or 0
if starting_id > max_id or ending_id > max_id:
return response_bad()
for i in range(starting_id, ending_id + 1, 1):
updated_bookmark = request.data.get(str(i)) # type: ignore
result_flag = bukudb.update_rec(
i,
updated_bookmark.get("url"),
updated_bookmark.get("title"),
updated_bookmark.get("tags"),
updated_bookmark.get("description"),
)
if result_flag is False:
return response_bad()
return response_ok()
def delete(self, starting_id: int, ending_id: int):
bukudb = getattr(flask.g, "bukudb", get_bukudb())
max_id = bukudb.get_max_id() or 0
if starting_id > max_id or ending_id > max_id:
return response_bad()
idx = min([starting_id, ending_id])
result_flag = bukudb.delete_rec(idx, starting_id, ending_id, is_range=True)
return to_response(result_flag)
class ApiBookmarkSearchView(MethodView):
def get(self):
arg_obj = request.args
keywords = arg_obj.getlist("keywords")
all_keywords = arg_obj.get("all_keywords")
deep = arg_obj.get("deep")
regex = arg_obj.get("regex")
# api request is more strict
all_keywords = False if all_keywords is None else all_keywords
deep = False if deep is None else deep
regex = False if regex is None else regex
all_keywords = (
all_keywords
if isinstance(all_keywords, bool)
else all_keywords.lower() == "true"
)
deep = deep if isinstance(deep, bool) else deep.lower() == "true"
regex = regex if isinstance(regex, bool) else regex.lower() == "true"
bukudb = getattr(flask.g, "bukudb", get_bukudb())
res = None
result = {
"bookmarks": [
entity(bookmark, id=True)
for bookmark in bukudb.searchdb(keywords, all_keywords, deep, regex)
]
}
current_app.logger.debug("total bookmarks:{}".format(len(result["bookmarks"])))
res = jsonify(result)
return res
def delete(self):
arg_obj = request.form
keywords = arg_obj.getlist("keywords")
all_keywords = arg_obj.get("all_keywords")
deep = arg_obj.get("deep")
regex = arg_obj.get("regex")
# api request is more strict
all_keywords = False if all_keywords is None else all_keywords
deep = False if deep is None else deep
regex = False if regex is None else regex
all_keywords = (
all_keywords
if isinstance(all_keywords, bool)
else all_keywords.lower() == "true"
)
deep = deep if isinstance(deep, bool) else deep.lower() == "true"
regex = regex if isinstance(regex, bool) else regex.lower() == "true"
bukudb = getattr(flask.g, "bukudb", get_bukudb())
res = None
for bookmark in bukudb.searchdb(keywords, all_keywords, deep, regex):
if not bukudb.delete_rec(bookmark.id):
res = response_bad()
return res or response_ok()
class BookmarkletView(MethodView): # pylint: disable=too-few-public-methods
def get(self):
url = request.args.get("url")
title = request.args.get("title")
description = request.args.get("description")
bukudb = getattr(flask.g, "bukudb", get_bukudb())
rec_id = bukudb.get_rec_id(url)
if rec_id:
return redirect(url_for("bookmark.edit_view", id=rec_id))
return redirect(
url_for(
"bookmark.create_view", link=url, title=title, description=description
)
)
|
sword | conftest | import os
import pytest
import xmlschema
from deposit.protocol import DepositResult
from deposit.sword.protocol import SWORDMETSMODSProtocol
from lxml import etree
@pytest.fixture
def sword_mods_protocol(request, repository):
"""
Creates a sword mods repository object
"""
sword_mods_repository = repository.sword_mods_repository()
request.cls.protocol = SWORDMETSMODSProtocol(sword_mods_repository)
@pytest.fixture()
def metadata_xml_dc():
"""
Returns bibliographic metadata as lxml etree. Use this fixture if you just need some metadata in XML format and it's content is not important.
"""
directory = os.path.dirname(os.path.abspath(__file__))
parser = etree.XMLParser(remove_blank_text=True)
return etree.parse(
os.path.join(directory, "test_data", "dc_lesebibliothek_frauenzimmer.xml"),
parser,
).getroot()
@pytest.fixture
def metadata_xml_mets():
"""
Returns a mets formatted xml with some metadata in dmdSec
"""
conftest_dir = os.path.dirname(os.path.abspath(__file__))
with open(
os.path.join(
conftest_dir, "test_data", "mets_dc_lesebibliothek_frauenzimmer.xml"
),
"r",
) as f:
xml = f.read()
return xml
@pytest.fixture(scope="session")
def validate_mets():
schema = xmlschema.XMLSchema(
"https://www.loc.gov/standards/mets/version112/mets.xsd"
)
def validator(xml):
schema.validate(xml)
return validator
@pytest.fixture(scope="session")
def validate_mods_3_7():
schema = xmlschema.XMLSchema("https://www.loc.gov/standards/mods/mods-3-7.xsd")
def validator(xml):
schema.validate(xml)
return validator
@pytest.fixture
def monkeypatch_metadata_creation(
request, monkeypatch, metadata_xml_dc, dissemin_xml_1_0
):
"""
Monkeypatches both functions, that generate metadata: _get_xml_metadata and _get_xml_dissemin_metadata
"""
monkeypatch.setattr(
request.cls.protocol, "_get_xml_dissemin_metadata", lambda x: dissemin_xml_1_0
)
monkeypatch.setattr(
request.cls.protocol, "_get_xml_metadata", lambda x: metadata_xml_dc
)
@pytest.fixture
def monkeypatch_get_deposit_result(request, monkeypatch):
"""
Mocks _get_deposit_reult so that it returns ``None`` and does not raise exception.
"""
def _get_deposit_result(*args, **kwargs):
deposit_result = DepositResult()
return deposit_result
monkeypatch.setattr(
request.cls.protocol, "_get_deposit_result", _get_deposit_result
)
|
events | squeezebox_sync | # Copyright 2011-2021 Nick Boultbee
#
# Inspired in parts by PySqueezeCenter (c) 2010 JingleManSweep
# SqueezeCenter and SqueezeBox are copyright Logitech
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
import os
from quodlibet import _, app, qltk
from quodlibet.ext._shared.squeezebox.base import SqueezeboxPluginMixin
from quodlibet.plugins.events import EventPlugin
from quodlibet.qltk import Icons
from quodlibet.util.dprint import print_d
if os.name == "nt":
from quodlibet.plugins import PluginNotSupportedError
raise PluginNotSupportedError
class SqueezeboxSyncPlugin(EventPlugin, SqueezeboxPluginMixin):
PLUGIN_ID = "Squeezebox Output"
PLUGIN_NAME = _("Squeezebox Sync")
PLUGIN_DESC_MARKUP = (
_(
"Makes Logitech Squeezebox mirror Quod Libet output, "
"provided both read from an identical library."
)
+ "\n"
+ _(
'Shares configuration with <a href="%(plugin_link)s">Export to '
"Squeezebox plugin</a>."
)
% {"plugin_link": "quodlibet:///prefs/plugins/Export to Squeezebox Playlist"}
)
PLUGIN_ICON = Icons.MEDIA_PLAYBACK_START
server = None
active = False
_debug = False
def __init__(self):
super().__init__()
@classmethod
def post_reconnect(cls):
cls.server.stop()
SqueezeboxPluginMixin.post_reconnect()
player = app.player
cls.plugin_on_song_started(player.info)
cls.plugin_on_seek(player.info, player.get_position())
def enabled(self):
print_d("Debug is set to %s" % self._debug)
self.active = True
self.init_server()
self.server.pause()
if not self.server.is_connected:
qltk.ErrorMessage(
None,
_("Error finding Squeezebox server"),
_("Error finding %s. Please check settings") % self.server.config,
).run()
def disabled(self):
# Stopping might be annoying in some situations, but seems more correct
if self.server:
self.server.stop()
self.active = False
@classmethod
def plugin_on_song_started(cls, song):
# Yucky hack to allow some form of immediacy on re-configuration
cls.server._debug = cls._debug = cls.config_get_bool("debug", False)
if cls._debug:
print_d("Paused" if app.player.paused else "Not paused")
if song and cls.server and cls.server.is_connected:
path = cls.get_sb_path(song)
print_d("Requesting to play %s..." % path)
if app.player.paused:
cls.server.change_song(path)
else:
cls.server.playlist_play(path)
@classmethod
def plugin_on_paused(cls):
if cls.server:
cls.server.pause()
@classmethod
def plugin_on_unpaused(cls):
if cls.server:
cls.server.unpause()
@classmethod
def plugin_on_seek(cls, song, msec):
if not app.player.paused:
if cls.server:
cls.server.seek_to(msec)
cls.server.play()
|
mainwin | patterncombo | import sys
from PyQt5.QtCore import Qt, pyqtSignal
from PyQt5.QtWidgets import (
QAbstractItemView,
QApplication,
QComboBox,
QFrame,
QHBoxLayout,
QInputDialog,
QPushButton,
QShortcut,
QVBoxLayout,
)
from ..puddleobjects import ListBox, ListButtons, PuddleConfig
from ..translations import translate
def load_patterns(filepath=None):
settings = PuddleConfig(filepath)
return [
settings.get(
"editor",
"patterns",
[
"%artist% - $num(%track%,2) - %title%",
"%artist% - %album% - $num(%track%,2) - %title%",
"%artist% - %title%",
"%artist% - %album%",
"%artist% - Track %track%",
"%artist% - %title%",
],
),
settings.get("editor", "index", 0, True),
]
class PatternCombo(QComboBox):
name = "toolbar-patterncombo"
patternchanged = pyqtSignal(str, name="patternchanged")
def __init__(self, items=None, parent=None, status=None):
self.emits = ["patternchanged"]
self.receives = [("patterns", self.setItems)]
self.settingsdialog = SettingsWin
QComboBox.__init__(self, parent)
self._status = status
status["patterns"] = self.items
status["patterntext"] = lambda: str(self.currentText())
self.setEditable(True)
if items:
self.addItems(items)
pchange = lambda text: self.patternchanged.emit(str(text))
self.editTextChanged.connect(pchange)
shortcut = QShortcut(self)
shortcut.setKey("F8")
shortcut.setContext(Qt.ShortcutContext.ApplicationShortcut)
def set_focus():
if self.hasFocus():
status["table"].setFocus()
else:
self.lineEdit().selectAll()
self.setFocus()
shortcut.activated.connect(set_focus)
def setItems(self, texts):
self.clear()
self.addItems(texts)
def items(self):
text = self.itemText
return [str(text(i)) for i in range(self.count())]
def loadSettings(self):
patterns, index = load_patterns()
self.setItems(patterns)
self.setCurrentIndex(index)
def saveSettings(self):
settings = PuddleConfig()
settings.set("editor", "patterns", list(self.items()))
settings.set("editor", "index", self.currentIndex())
class SettingsWin(QFrame):
def __init__(self, parent=None, cenwid=None, status=None):
QFrame.__init__(self, parent)
self.title = translate("Settings", "Patterns")
connect = lambda c, signal, s: getattr(c, signal).connect(s)
self.setFrameStyle(QFrame.Shape.Box)
self.listbox = ListBox()
self.listbox.setSelectionMode(QAbstractItemView.SelectionMode.ExtendedSelection)
buttons = ListButtons()
self.listbox.addItems(status["patterns"])
hbox = QHBoxLayout()
hbox.addWidget(self.listbox)
self.setLayout(hbox)
vbox = QVBoxLayout()
sortlistbox = QPushButton(translate("Pattern Settings", "&Sort"))
self._sortOrder = Qt.SortOrder.AscendingOrder
connect(sortlistbox, "clicked", self._sortListBox)
vbox.addWidget(sortlistbox)
vbox.addLayout(buttons)
vbox.addStretch()
hbox.addLayout(vbox)
connect(buttons, "add", self.addPattern)
connect(buttons, "edit", self.editItem)
buttons.duplicateButton.setVisible(False)
self.listbox.connectToListButtons(buttons)
self.listbox.editButton = buttons.editButton
connect(self.listbox, "itemDoubleClicked", self._doubleClicked)
def _sortListBox(self):
if self._sortOrder == Qt.SortOrder.AscendingOrder:
self.listbox.sortItems(Qt.SortOrder.DescendingOrder)
self._sortOrder = Qt.SortOrder.DescendingOrder
else:
self.listbox.sortItems(Qt.SortOrder.AscendingOrder)
self._sortOrder = Qt.SortOrder.AscendingOrder
def saveSettings(self):
patterns = [
str(self.listbox.item(row).text()) for row in range(self.listbox.count())
]
cparser = PuddleConfig()
cparser.setSection("editor", "patterns", patterns)
def addPattern(self):
l = self.listbox.item
patterns = [str(l(z).text()) for z in range(self.listbox.count())]
row = self.listbox.currentRow()
if row < 0:
row = 0
(text, ok) = QInputDialog().getItem(
self,
"puddletag",
translate("Pattern Settings", "Enter a pattern"),
patterns,
row,
)
if ok:
self.listbox.clearSelection()
self.listbox.addItem(text)
self.listbox.setCurrentRow(self.listbox.count() - 1)
def _doubleClicked(self, item):
self.editItem()
def editItem(self, row=None):
if row is None:
row = self.listbox.currentRow()
l = self.listbox.item
patterns = [str(l(z).text()) for z in range(self.listbox.count())]
(text, ok) = QInputDialog().getItem(
self,
"puddletag",
translate("Pattern Settings", "Enter a pattern"),
patterns,
row,
)
if ok:
item = l(row)
item.setText(text)
item.setSelected(True)
def applySettings(self, control):
item = self.listbox.item
patterns = [item(row).text() for row in range(self.listbox.count())]
control.setItems(patterns)
control = ("patterncombo", PatternCombo, False)
if __name__ == "__main__":
app = QApplication(sys.argv)
win = PatternCombo(["one", "the", "three"])
win.show()
app.exec_()
|
Cloud | CloudApiClient | # Copyright (c) 2019 Ultimaker B.V.
# Cura is released under the terms of the LGPLv3 or higher.
import json
import urllib.parse
from json import JSONDecodeError
from time import time
from typing import (
Any,
Callable,
Dict,
List,
Optional,
Tuple,
Type,
TypeVar,
Union,
cast,
)
from cura.API import Account
from cura.CuraApplication import CuraApplication
from cura.UltimakerCloud import UltimakerCloudConstants
from cura.UltimakerCloud.UltimakerCloudScope import UltimakerCloudScope
from PyQt6.QtCore import QUrl
from PyQt6.QtNetwork import QNetworkReply, QNetworkRequest
from UM.Logger import Logger
from UM.TaskManagement.HttpRequestManager import HttpRequestManager
from UM.TaskManagement.HttpRequestScope import JsonDecoratorScope
from ..Models.BaseModel import BaseModel
from ..Models.Http.CloudClusterResponse import CloudClusterResponse
from ..Models.Http.CloudClusterStatus import CloudClusterStatus
from ..Models.Http.CloudClusterWithConfigResponse import CloudClusterWithConfigResponse
from ..Models.Http.CloudError import CloudError
from ..Models.Http.CloudPrintJobResponse import CloudPrintJobResponse
from ..Models.Http.CloudPrintJobUploadRequest import CloudPrintJobUploadRequest
from ..Models.Http.CloudPrintResponse import CloudPrintResponse
from .ToolPathUploader import ToolPathUploader
CloudApiClientModel = TypeVar("CloudApiClientModel", bound=BaseModel)
"""The generic type variable used to document the methods below."""
class CloudApiClient:
"""The cloud API client is responsible for handling the requests and responses from the cloud.
Each method should only handle models instead of exposing Any HTTP details.
"""
# The cloud URL to use for this remote cluster.
ROOT_PATH = UltimakerCloudConstants.CuraCloudAPIRoot
CLUSTER_API_ROOT = "{}/connect/v1".format(ROOT_PATH)
CURA_API_ROOT = "{}/cura/v1".format(ROOT_PATH)
DEFAULT_REQUEST_TIMEOUT = 10 # seconds
# In order to avoid garbage collection we keep the callbacks in this list.
_anti_gc_callbacks = [] # type: List[Callable[[Any], None]]
def __init__(
self, app: CuraApplication, on_error: Callable[[List[CloudError]], None]
) -> None:
"""Initializes a new cloud API client.
:param app:
:param on_error: The callback to be called whenever we receive errors from the server.
"""
super().__init__()
self._app = app
self._account = app.getCuraAPI().account
self._scope = JsonDecoratorScope(UltimakerCloudScope(app))
self._http = HttpRequestManager.getInstance()
self._on_error = on_error
self._upload: Optional[ToolPathUploader] = None
@property
def account(self) -> Account:
"""Gets the account used for the API."""
return self._account
def getClusters(
self, on_finished: Callable[[List[CloudClusterResponse]], Any], failed: Callable
) -> None:
"""Retrieves all the clusters for the user that is currently logged in.
:param on_finished: The function to be called after the result is parsed.
"""
url = f"{self.CLUSTER_API_ROOT}/clusters?status=active"
self._http.get(
url,
scope=self._scope,
callback=self._parseCallback(on_finished, CloudClusterResponse, failed),
error_callback=failed,
timeout=self.DEFAULT_REQUEST_TIMEOUT,
)
def getClustersByMachineType(
self,
machine_type,
on_finished: Callable[[List[CloudClusterWithConfigResponse]], Any],
failed: Callable,
) -> None:
# HACK: There is something weird going on with the API, as it reports printer types in formats like
# "ultimaker_s3", but wants "Ultimaker S3" when using the machine_variant filter query. So we need to do some
# conversion!
machine_type = machine_type.replace("_plus", "+")
machine_type = machine_type.replace("_", " ")
machine_type = machine_type.replace("ultimaker", "ultimaker ")
machine_type = machine_type.replace(" ", " ")
machine_type = machine_type.title()
machine_type = urllib.parse.quote_plus(machine_type)
url = f"{self.CLUSTER_API_ROOT}/clusters?machine_variant={machine_type}"
self._http.get(
url,
scope=self._scope,
callback=self._parseCallback(
on_finished, CloudClusterWithConfigResponse, failed
),
error_callback=failed,
timeout=self.DEFAULT_REQUEST_TIMEOUT,
)
def getClusterStatus(
self, cluster_id: str, on_finished: Callable[[CloudClusterStatus], Any]
) -> None:
"""Retrieves the status of the given cluster.
:param cluster_id: The ID of the cluster.
:param on_finished: The function to be called after the result is parsed.
"""
url = f"{self.CLUSTER_API_ROOT}/clusters/{cluster_id}/status"
self._http.get(
url,
scope=self._scope,
callback=self._parseCallback(on_finished, CloudClusterStatus),
timeout=self.DEFAULT_REQUEST_TIMEOUT,
)
def requestUpload(
self,
request: CloudPrintJobUploadRequest,
on_finished: Callable[[CloudPrintJobResponse], Any],
) -> None:
"""Requests the cloud to register the upload of a print job mesh.
:param request: The request object.
:param on_finished: The function to be called after the result is parsed.
"""
url = f"{self.CURA_API_ROOT}/jobs/upload"
data = json.dumps({"data": request.toDict()}).encode()
self._http.put(
url,
scope=self._scope,
data=data,
callback=self._parseCallback(on_finished, CloudPrintJobResponse),
timeout=self.DEFAULT_REQUEST_TIMEOUT,
)
def uploadToolPath(
self,
print_job: CloudPrintJobResponse,
mesh: bytes,
on_finished: Callable[[], Any],
on_progress: Callable[[int], Any],
on_error: Callable[[], Any],
):
"""Uploads a print job tool path to the cloud.
:param print_job: The object received after requesting an upload with `self.requestUpload`.
:param mesh: The tool path data to be uploaded.
:param on_finished: The function to be called after the upload is successful.
:param on_progress: A function to be called during upload progress. It receives a percentage (0-100).
:param on_error: A function to be called if the upload fails.
"""
self._upload = ToolPathUploader(
self._http, print_job, mesh, on_finished, on_progress, on_error
)
self._upload.start()
# Requests a cluster to print the given print job.
# \param cluster_id: The ID of the cluster.
# \param job_id: The ID of the print job.
# \param on_finished: The function to be called after the result is parsed.
# \param on_error: A function to be called if there was a server-side problem uploading. Generic errors (not
# specific to sending print jobs) such as lost connection, unparsable responses, etc. are not returned here, but
# handled in a generic way by the CloudApiClient.
def requestPrint(
self,
cluster_id: str,
job_id: str,
on_finished: Callable[[CloudPrintResponse], Any],
on_error,
) -> None:
url = f"{self.CLUSTER_API_ROOT}/clusters/{cluster_id}/print/{job_id}"
self._http.post(
url,
scope=self._scope,
data=b"",
callback=self._parseCallback(on_finished, CloudPrintResponse),
error_callback=on_error,
timeout=self.DEFAULT_REQUEST_TIMEOUT,
)
def doPrintJobAction(
self,
cluster_id: str,
cluster_job_id: str,
action: str,
data: Optional[Dict[str, Any]] = None,
) -> None:
"""Send a print job action to the cluster for the given print job.
:param cluster_id: The ID of the cluster.
:param cluster_job_id: The ID of the print job within the cluster.
:param action: The name of the action to execute.
"""
body = json.dumps({"data": data}).encode() if data else b""
url = f"{self.CLUSTER_API_ROOT}/clusters/{cluster_id}/print_jobs/{cluster_job_id}/action/{action}"
self._http.post(
url, scope=self._scope, data=body, timeout=self.DEFAULT_REQUEST_TIMEOUT
)
def _createEmptyRequest(
self, path: str, content_type: Optional[str] = "application/json"
) -> QNetworkRequest:
"""We override _createEmptyRequest in order to add the user credentials.
:param path: The URL to request
:param content_type: The type of the body contents.
"""
request = QNetworkRequest(QUrl(path))
if content_type:
request.setHeader(
QNetworkRequest.KnownHeaders.ContentTypeHeader, content_type
)
access_token = self._account.accessToken
if access_token:
request.setRawHeader(b"Authorization", f"Bearer {access_token}".encode())
return request
@staticmethod
def _parseReply(reply: QNetworkReply) -> Tuple[int, Dict[str, Any]]:
"""Parses the given JSON network reply into a status code and a dictionary, handling unexpected errors as well.
:param reply: The reply from the server.
:return: A tuple with a status code and a dictionary.
"""
status_code = reply.attribute(QNetworkRequest.Attribute.HttpStatusCodeAttribute)
try:
response = bytes(reply.readAll()).decode()
return status_code, json.loads(response)
except (UnicodeDecodeError, JSONDecodeError, ValueError) as err:
error = CloudError(
code=type(err).__name__,
title=str(err),
http_code=str(status_code),
id=str(time()),
http_status="500",
)
Logger.logException(
"e", "Could not parse the stardust response: %s", error.toDict()
)
return status_code, {"errors": [error.toDict()]}
def _parseResponse(
self,
response: Dict[str, Any],
on_finished: Union[
Callable[[CloudApiClientModel], Any],
Callable[[List[CloudApiClientModel]], Any],
],
model_class: Type[CloudApiClientModel],
) -> None:
"""Parses the given response and calls the correct callback depending on the result.
:param response: The response from the server, after being converted to a dict.
:param on_finished: The callback in case the response is successful.
:param model_class: The type of the model to convert the response to. It may either be a single record or a list.
"""
if "data" in response:
data = response["data"]
if "status" in data and data["status"] == "wait_approval":
on_finished_empty = cast(Callable[[List], Any], on_finished)
on_finished_empty([])
elif isinstance(data, list):
results = [model_class(**c) for c in data] # type: List[CloudApiClientModel]
on_finished_list = cast(
Callable[[List[CloudApiClientModel]], Any], on_finished
)
on_finished_list(results)
else:
result = model_class(**data) # type: CloudApiClientModel
on_finished_item = cast(
Callable[[CloudApiClientModel], Any], on_finished
)
on_finished_item(result)
elif "errors" in response:
self._on_error([CloudError(**error) for error in response["errors"]])
else:
Logger.log(
"e", "Cannot find data or errors in the cloud response: %s", response
)
def _parseCallback(
self,
on_finished: Union[
Callable[[CloudApiClientModel], Any],
Callable[[List[CloudApiClientModel]], Any],
],
model: Type[CloudApiClientModel],
on_error: Optional[Callable] = None,
) -> Callable[[QNetworkReply], None]:
"""Creates a callback function so that it includes the parsing of the response into the correct model.
The callback is added to the 'finished' signal of the reply.
:param reply: The reply that should be listened to.
:param on_finished: The callback in case the response is successful. Depending on the endpoint it will be either
a list or a single item.
:param model: The type of the model to convert the response to.
"""
def parse(reply: QNetworkReply) -> None:
self._anti_gc_callbacks.remove(parse)
# Don't try to parse the reply if we didn't get one
if (
reply.attribute(QNetworkRequest.Attribute.HttpStatusCodeAttribute)
is None
):
if on_error is not None:
on_error()
return
status_code, response = self._parseReply(reply)
if status_code >= 300 and on_error is not None:
on_error()
else:
self._parseResponse(response, on_finished, model)
self._anti_gc_callbacks.append(parse)
return parse
|
PythonBrowser | PythonBrowserModel | """PythonBrowserModel.py -- module implementing the data model for PythonBrowser."""
import sys
from operator import getitem, setitem
from AppKit import NSBeep
from Foundation import NSObject
class PythonBrowserModel(NSObject):
"""This is a delegate as well as a data source for NSOutlineViews."""
def initWithObject_(self, obj):
self = self.init()
self.setObject_(obj)
return self
def setObject_(self, obj):
self.root = PythonItem("<root>", obj, None, None)
# NSOutlineViewDataSource methods
def outlineView_numberOfChildrenOfItem_(self, view, item):
if item is None:
item = self.root
return len(item)
def outlineView_child_ofItem_(self, view, child, item):
if item is None:
item = self.root
return item.getChild(child)
def outlineView_isItemExpandable_(self, view, item):
if item is None:
item = self.root
return item.isExpandable()
def outlineView_objectValueForTableColumn_byItem_(self, view, col, item):
if item is None:
item = self.root
return getattr(item, col.identifier())
def outlineView_setObjectValue_forTableColumn_byItem_(self, view, value, col, item):
assert col.identifier() == "value"
if item.value == value:
return
try:
obj = eval(value, {})
except:
NSBeep()
print "XXX Error:", sys.exc_info()
print "XXX :", repr(value)
else:
item.setValue(obj)
# delegate method
def outlineView_shouldEditTableColumn_item_(self, view, col, item):
return item.isEditable()
# objects of these types are not eligable for expansion in the outline view
SIMPLE_TYPES = (str, unicode, int, long, float, complex)
def getInstanceVarNames(obj):
"""Return a list the names of all (potential) instance variables."""
# Recipe from Guido
slots = {}
if hasattr(obj, "__dict__"):
slots.update(obj.__dict__)
if hasattr(obj, "__class__"):
slots["__class__"] = 1
cls = getattr(obj, "__class__", type(obj))
if hasattr(cls, "__mro__"):
for base in cls.__mro__:
for name, value in base.__dict__.items():
# XXX using callable() is a heuristic which isn't 100%
# foolproof.
if hasattr(value, "__get__") and not callable(value) and \
hasattr(obj, name):
slots[name] = 1
if "__dict__" in slots:
del slots["__dict__"]
slots = slots.keys()
slots.sort()
return slots
class NiceError:
"""Wrapper for an exception so we can display it nicely in the browser."""
def __init__(self, exc_info):
self.exc_info = exc_info
def __repr__(self):
from traceback import format_exception_only
lines = format_exception_only(*self.exc_info[:2])
assert len(lines) == 1
error = lines[0].strip()
return "*** error *** %s" %error
class PythonItem(NSObject):
"""Wrapper class for items to be displayed in the outline view."""
# We keep references to all child items (once created). This is
# neccesary because NSOutlineView holds on to PythonItem instances
# without retaining them. If we don't make sure they don't get
# garbage collected, the app will crash. For the same reason this
# class _must_ derive from NSObject, since otherwise autoreleased
# proxies will be fed to NSOutlineView, which will go away too soon.
def __new__(cls, *args, **kwargs):
# "Pythonic" constructor
return cls.alloc().init()
def __init__(self, name, obj, parent, setvalue):
self.realName = name
self.name = str(name)
self.parent = parent
self._setValue = setvalue
self.type = type(obj).__name__
try:
self.value = repr(obj)[:256] # XXX [:256] makes it quite a bit faster for long reprs.
assert isinstance(self.value, str)
except:
self.value = repr(NiceError(sys.exc_info()))
self.object = obj
self.childrenEditable = 0
if isinstance(obj, dict):
self.children = obj.keys()
self.children.sort()
self._getChild = getitem
self._setChild = setitem
self.childrenEditable = 1
elif obj is None or isinstance(obj, SIMPLE_TYPES):
self._getChild = None
self._setChild = None
elif isinstance(obj, (list, tuple)):
self.children = range(len(obj))
self._getChild = getitem
self._setChild = setitem
if isinstance(obj, list):
self.childrenEditable = 1
else:
self.children = getInstanceVarNames(obj)
self._getChild = getattr
self._setChild = setattr
self.childrenEditable = 1 # XXX we don't know that...
self._childRefs = {}
def setValue(self, value):
self._setValue(self.parent, self.realName, value)
self.__init__(self.realName, value, self.parent, self._setValue)
def isEditable(self):
return self._setValue is not None
def isExpandable(self):
return self._getChild is not None
def getChild(self, child):
if self._childRefs.has_key(child):
return self._childRefs[child]
name = self.children[child]
try:
obj = self._getChild(self.object, name)
except:
obj = NiceError(sys.exc_info())
if self.childrenEditable:
childObj = PythonItem(name, obj, self.object, self._setChild)
else:
childObj = PythonItem(name, obj, None, None)
self._childRefs[child] = childObj
return childObj
def __len__(self):
return len(self.children)
|
views | telegram_channels | from apps.api.permissions import RBACPermission
from apps.api.serializers.telegram import TelegramToOrganizationConnectorSerializer
from apps.auth_token.auth import PluginAuthentication
from common.api_helpers.mixins import PublicPrimaryKeyMixin
from common.insight_log.chatops_insight_logs import (
ChatOpsEvent,
ChatOpsTypePlug,
write_chatops_insight_log,
)
from rest_framework import mixins, status, viewsets
from rest_framework.decorators import action
from rest_framework.permissions import IsAuthenticated
from rest_framework.response import Response
class TelegramChannelViewSet(
PublicPrimaryKeyMixin,
mixins.RetrieveModelMixin,
mixins.DestroyModelMixin,
mixins.ListModelMixin,
viewsets.GenericViewSet,
):
authentication_classes = (PluginAuthentication,)
permission_classes = (IsAuthenticated, RBACPermission)
rbac_permissions = {
"metadata": [RBACPermission.Permissions.CHATOPS_READ],
"list": [RBACPermission.Permissions.CHATOPS_READ],
"retrieve": [RBACPermission.Permissions.CHATOPS_READ],
"destroy": [RBACPermission.Permissions.CHATOPS_UPDATE_SETTINGS],
"set_default": [RBACPermission.Permissions.CHATOPS_UPDATE_SETTINGS],
}
serializer_class = TelegramToOrganizationConnectorSerializer
def get_queryset(self):
from apps.telegram.models import TelegramToOrganizationConnector
return TelegramToOrganizationConnector.objects.filter(
organization=self.request.user.organization
)
@action(detail=True, methods=["post"])
def set_default(self, request, pk):
telegram_channel = self.get_object()
telegram_channel.make_channel_default(request.user)
return Response(status=status.HTTP_200_OK)
def perform_destroy(self, instance):
user = self.request.user
write_chatops_insight_log(
author=user,
event_name=ChatOpsEvent.CHANNEL_DISCONNECTED,
chatops_type=ChatOpsTypePlug.TELEGRAM.value,
channel_name=instance.channel_name,
)
instance.delete()
|
jargon | resources | # =============================================================================
# Copyright (C) 2018 Filip Sufitchi
#
# This file is part of pyfa.
#
# pyfa is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# pyfa is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with pyfa. If not, see <http://www.gnu.org/licenses/>.
# =============================================================================
import pkg_resources
DEFAULT_DATA = pkg_resources.resource_string(__name__, "defaults.yaml").decode()
DEFAULT_HEADER = pkg_resources.resource_string(__name__, "header.yaml").decode()
|
generated | indexsuper | #!/usr/bin/env python
#
# Generated Thu Jun 11 18:43:54 2009 by generateDS.py.
#
import sys
from xml.dom import Node, minidom
#
# User methods
#
# Calls to the methods in these classes are generated by generateDS.py.
# You can replace these methods by re-implementing the following class
# in a module named generatedssuper.py.
try:
from generatedssuper import GeneratedsSuper
except ImportError as exp:
class GeneratedsSuper(object):
def format_string(self, input_data, input_name=""):
return input_data
def format_integer(self, input_data, input_name=""):
return "%d" % input_data
def format_float(self, input_data, input_name=""):
return "%f" % input_data
def format_double(self, input_data, input_name=""):
return "%e" % input_data
def format_boolean(self, input_data, input_name=""):
return "%s" % input_data
#
# If you have installed IPython you can uncomment and use the following.
# IPython is available from http://ipython.scipy.org/.
#
## from IPython.Shell import IPShellEmbed
## args = ''
# ipshell = IPShellEmbed(args,
## banner = 'Dropping into IPython',
# exit_msg = 'Leaving Interpreter, back to program.')
# Then use the following line where and when you want to drop into the
# IPython shell:
# ipshell('<some message> -- Entering ipshell.\nHit Ctrl-D to exit')
#
# Globals
#
ExternalEncoding = "ascii"
#
# Support/utility functions.
#
def showIndent(outfile, level):
for idx in range(level):
outfile.write(" ")
def quote_xml(inStr):
s1 = isinstance(inStr, str) and inStr or "%s" % inStr
s1 = s1.replace("&", "&")
s1 = s1.replace("<", "<")
s1 = s1.replace(">", ">")
return s1
def quote_attrib(inStr):
s1 = isinstance(inStr, str) and inStr or "%s" % inStr
s1 = s1.replace("&", "&")
s1 = s1.replace("<", "<")
s1 = s1.replace(">", ">")
if '"' in s1:
if "'" in s1:
s1 = '"%s"' % s1.replace('"', """)
else:
s1 = "'%s'" % s1
else:
s1 = '"%s"' % s1
return s1
def quote_python(inStr):
s1 = inStr
if s1.find("'") == -1:
if s1.find("\n") == -1:
return "'%s'" % s1
else:
return "'''%s'''" % s1
else:
if s1.find('"') != -1:
s1 = s1.replace('"', '\\"')
if s1.find("\n") == -1:
return '"%s"' % s1
else:
return '"""%s"""' % s1
class MixedContainer(object):
# Constants for category:
CategoryNone = 0
CategoryText = 1
CategorySimple = 2
CategoryComplex = 3
# Constants for content_type:
TypeNone = 0
TypeText = 1
TypeString = 2
TypeInteger = 3
TypeFloat = 4
TypeDecimal = 5
TypeDouble = 6
TypeBoolean = 7
def __init__(self, category, content_type, name, value):
self.category = category
self.content_type = content_type
self.name = name
self.value = value
def getCategory(self):
return self.category
def getContenttype(self, content_type):
return self.content_type
def getValue(self):
return self.value
def getName(self):
return self.name
def export(self, outfile, level, name, namespace):
if self.category == MixedContainer.CategoryText:
outfile.write(self.value)
elif self.category == MixedContainer.CategorySimple:
self.exportSimple(outfile, level, name)
else: # category == MixedContainer.CategoryComplex
self.value.export(outfile, level, namespace, name)
def exportSimple(self, outfile, level, name):
if self.content_type == MixedContainer.TypeString:
outfile.write("<%s>%s</%s>" % (self.name, self.value, self.name))
elif (
self.content_type == MixedContainer.TypeInteger
or self.content_type == MixedContainer.TypeBoolean
):
outfile.write("<%s>%d</%s>" % (self.name, self.value, self.name))
elif (
self.content_type == MixedContainer.TypeFloat
or self.content_type == MixedContainer.TypeDecimal
):
outfile.write("<%s>%f</%s>" % (self.name, self.value, self.name))
elif self.content_type == MixedContainer.TypeDouble:
outfile.write("<%s>%g</%s>" % (self.name, self.value, self.name))
def exportLiteral(self, outfile, level, name):
if self.category == MixedContainer.CategoryText:
showIndent(outfile, level)
outfile.write(
'MixedContainer(%d, %d, "%s", "%s"),\n'
% (self.category, self.content_type, self.name, self.value)
)
elif self.category == MixedContainer.CategorySimple:
showIndent(outfile, level)
outfile.write(
'MixedContainer(%d, %d, "%s", "%s"),\n'
% (self.category, self.content_type, self.name, self.value)
)
else: # category == MixedContainer.CategoryComplex
showIndent(outfile, level)
outfile.write(
'MixedContainer(%d, %d, "%s",\n'
% (
self.category,
self.content_type,
self.name,
)
)
self.value.exportLiteral(outfile, level + 1)
showIndent(outfile, level)
outfile.write(")\n")
class _MemberSpec(object):
def __init__(self, name="", data_type="", container=0):
self.name = name
self.data_type = data_type
self.container = container
def set_name(self, name):
self.name = name
def get_name(self):
return self.name
def set_data_type(self, data_type):
self.data_type = data_type
def get_data_type(self):
return self.data_type
def set_container(self, container):
self.container = container
def get_container(self):
return self.container
#
# Data representation classes.
#
class DoxygenType(GeneratedsSuper):
subclass = None
superclass = None
def __init__(self, version=None, compound=None):
self.version = version
if compound is None:
self.compound = []
else:
self.compound = compound
def factory(*args_, **kwargs_):
if DoxygenType.subclass:
return DoxygenType.subclass(*args_, **kwargs_)
else:
return DoxygenType(*args_, **kwargs_)
factory = staticmethod(factory)
def get_compound(self):
return self.compound
def set_compound(self, compound):
self.compound = compound
def add_compound(self, value):
self.compound.append(value)
def insert_compound(self, index, value):
self.compound[index] = value
def get_version(self):
return self.version
def set_version(self, version):
self.version = version
def export(
self, outfile, level, namespace_="", name_="DoxygenType", namespacedef_=""
):
showIndent(outfile, level)
outfile.write(
"<%s%s %s"
% (
namespace_,
name_,
namespacedef_,
)
)
self.exportAttributes(outfile, level, namespace_, name_="DoxygenType")
if self.hasContent_():
outfile.write(">\n")
self.exportChildren(outfile, level + 1, namespace_, name_)
showIndent(outfile, level)
outfile.write("</%s%s>\n" % (namespace_, name_))
else:
outfile.write(" />\n")
def exportAttributes(self, outfile, level, namespace_="", name_="DoxygenType"):
outfile.write(
" version=%s"
% (
self.format_string(
quote_attrib(self.version).encode(ExternalEncoding),
input_name="version",
),
)
)
def exportChildren(self, outfile, level, namespace_="", name_="DoxygenType"):
for compound_ in self.compound:
compound_.export(outfile, level, namespace_, name_="compound")
def hasContent_(self):
if self.compound is not None:
return True
else:
return False
def exportLiteral(self, outfile, level, name_="DoxygenType"):
level += 1
self.exportLiteralAttributes(outfile, level, name_)
if self.hasContent_():
self.exportLiteralChildren(outfile, level, name_)
def exportLiteralAttributes(self, outfile, level, name_):
if self.version is not None:
showIndent(outfile, level)
outfile.write("version = %s,\n" % (self.version,))
def exportLiteralChildren(self, outfile, level, name_):
showIndent(outfile, level)
outfile.write("compound=[\n")
level += 1
for compound in self.compound:
showIndent(outfile, level)
outfile.write("model_.compound(\n")
compound.exportLiteral(outfile, level, name_="compound")
showIndent(outfile, level)
outfile.write("),\n")
level -= 1
showIndent(outfile, level)
outfile.write("],\n")
def build(self, node_):
attrs = node_.attributes
self.buildAttributes(attrs)
for child_ in node_.childNodes:
nodeName_ = child_.nodeName.split(":")[-1]
self.buildChildren(child_, nodeName_)
def buildAttributes(self, attrs):
if attrs.get("version"):
self.version = attrs.get("version").value
def buildChildren(self, child_, nodeName_):
if child_.nodeType == Node.ELEMENT_NODE and nodeName_ == "compound":
obj_ = CompoundType.factory()
obj_.build(child_)
self.compound.append(obj_)
# end class DoxygenType
class CompoundType(GeneratedsSuper):
subclass = None
superclass = None
def __init__(self, kind=None, refid=None, name=None, member=None):
self.kind = kind
self.refid = refid
self.name = name
if member is None:
self.member = []
else:
self.member = member
def factory(*args_, **kwargs_):
if CompoundType.subclass:
return CompoundType.subclass(*args_, **kwargs_)
else:
return CompoundType(*args_, **kwargs_)
factory = staticmethod(factory)
def get_name(self):
return self.name
def set_name(self, name):
self.name = name
def get_member(self):
return self.member
def set_member(self, member):
self.member = member
def add_member(self, value):
self.member.append(value)
def insert_member(self, index, value):
self.member[index] = value
def get_kind(self):
return self.kind
def set_kind(self, kind):
self.kind = kind
def get_refid(self):
return self.refid
def set_refid(self, refid):
self.refid = refid
def export(
self, outfile, level, namespace_="", name_="CompoundType", namespacedef_=""
):
showIndent(outfile, level)
outfile.write(
"<%s%s %s"
% (
namespace_,
name_,
namespacedef_,
)
)
self.exportAttributes(outfile, level, namespace_, name_="CompoundType")
if self.hasContent_():
outfile.write(">\n")
self.exportChildren(outfile, level + 1, namespace_, name_)
showIndent(outfile, level)
outfile.write("</%s%s>\n" % (namespace_, name_))
else:
outfile.write(" />\n")
def exportAttributes(self, outfile, level, namespace_="", name_="CompoundType"):
outfile.write(" kind=%s" % (quote_attrib(self.kind),))
outfile.write(
" refid=%s"
% (
self.format_string(
quote_attrib(self.refid).encode(ExternalEncoding),
input_name="refid",
),
)
)
def exportChildren(self, outfile, level, namespace_="", name_="CompoundType"):
if self.name is not None:
showIndent(outfile, level)
outfile.write(
"<%sname>%s</%sname>\n"
% (
namespace_,
self.format_string(
quote_xml(self.name).encode(ExternalEncoding), input_name="name"
),
namespace_,
)
)
for member_ in self.member:
member_.export(outfile, level, namespace_, name_="member")
def hasContent_(self):
if self.name is not None or self.member is not None:
return True
else:
return False
def exportLiteral(self, outfile, level, name_="CompoundType"):
level += 1
self.exportLiteralAttributes(outfile, level, name_)
if self.hasContent_():
self.exportLiteralChildren(outfile, level, name_)
def exportLiteralAttributes(self, outfile, level, name_):
if self.kind is not None:
showIndent(outfile, level)
outfile.write('kind = "%s",\n' % (self.kind,))
if self.refid is not None:
showIndent(outfile, level)
outfile.write("refid = %s,\n" % (self.refid,))
def exportLiteralChildren(self, outfile, level, name_):
showIndent(outfile, level)
outfile.write("name=%s,\n" % quote_python(self.name).encode(ExternalEncoding))
showIndent(outfile, level)
outfile.write("member=[\n")
level += 1
for member in self.member:
showIndent(outfile, level)
outfile.write("model_.member(\n")
member.exportLiteral(outfile, level, name_="member")
showIndent(outfile, level)
outfile.write("),\n")
level -= 1
showIndent(outfile, level)
outfile.write("],\n")
def build(self, node_):
attrs = node_.attributes
self.buildAttributes(attrs)
for child_ in node_.childNodes:
nodeName_ = child_.nodeName.split(":")[-1]
self.buildChildren(child_, nodeName_)
def buildAttributes(self, attrs):
if attrs.get("kind"):
self.kind = attrs.get("kind").value
if attrs.get("refid"):
self.refid = attrs.get("refid").value
def buildChildren(self, child_, nodeName_):
if child_.nodeType == Node.ELEMENT_NODE and nodeName_ == "name":
name_ = ""
for text__content_ in child_.childNodes:
name_ += text__content_.nodeValue
self.name = name_
elif child_.nodeType == Node.ELEMENT_NODE and nodeName_ == "member":
obj_ = MemberType.factory()
obj_.build(child_)
self.member.append(obj_)
# end class CompoundType
class MemberType(GeneratedsSuper):
subclass = None
superclass = None
def __init__(self, kind=None, refid=None, name=None):
self.kind = kind
self.refid = refid
self.name = name
def factory(*args_, **kwargs_):
if MemberType.subclass:
return MemberType.subclass(*args_, **kwargs_)
else:
return MemberType(*args_, **kwargs_)
factory = staticmethod(factory)
def get_name(self):
return self.name
def set_name(self, name):
self.name = name
def get_kind(self):
return self.kind
def set_kind(self, kind):
self.kind = kind
def get_refid(self):
return self.refid
def set_refid(self, refid):
self.refid = refid
def export(
self, outfile, level, namespace_="", name_="MemberType", namespacedef_=""
):
showIndent(outfile, level)
outfile.write(
"<%s%s %s"
% (
namespace_,
name_,
namespacedef_,
)
)
self.exportAttributes(outfile, level, namespace_, name_="MemberType")
if self.hasContent_():
outfile.write(">\n")
self.exportChildren(outfile, level + 1, namespace_, name_)
showIndent(outfile, level)
outfile.write("</%s%s>\n" % (namespace_, name_))
else:
outfile.write(" />\n")
def exportAttributes(self, outfile, level, namespace_="", name_="MemberType"):
outfile.write(" kind=%s" % (quote_attrib(self.kind),))
outfile.write(
" refid=%s"
% (
self.format_string(
quote_attrib(self.refid).encode(ExternalEncoding),
input_name="refid",
),
)
)
def exportChildren(self, outfile, level, namespace_="", name_="MemberType"):
if self.name is not None:
showIndent(outfile, level)
outfile.write(
"<%sname>%s</%sname>\n"
% (
namespace_,
self.format_string(
quote_xml(self.name).encode(ExternalEncoding), input_name="name"
),
namespace_,
)
)
def hasContent_(self):
if self.name is not None:
return True
else:
return False
def exportLiteral(self, outfile, level, name_="MemberType"):
level += 1
self.exportLiteralAttributes(outfile, level, name_)
if self.hasContent_():
self.exportLiteralChildren(outfile, level, name_)
def exportLiteralAttributes(self, outfile, level, name_):
if self.kind is not None:
showIndent(outfile, level)
outfile.write('kind = "%s",\n' % (self.kind,))
if self.refid is not None:
showIndent(outfile, level)
outfile.write("refid = %s,\n" % (self.refid,))
def exportLiteralChildren(self, outfile, level, name_):
showIndent(outfile, level)
outfile.write("name=%s,\n" % quote_python(self.name).encode(ExternalEncoding))
def build(self, node_):
attrs = node_.attributes
self.buildAttributes(attrs)
for child_ in node_.childNodes:
nodeName_ = child_.nodeName.split(":")[-1]
self.buildChildren(child_, nodeName_)
def buildAttributes(self, attrs):
if attrs.get("kind"):
self.kind = attrs.get("kind").value
if attrs.get("refid"):
self.refid = attrs.get("refid").value
def buildChildren(self, child_, nodeName_):
if child_.nodeType == Node.ELEMENT_NODE and nodeName_ == "name":
name_ = ""
for text__content_ in child_.childNodes:
name_ += text__content_.nodeValue
self.name = name_
# end class MemberType
USAGE_TEXT = """
Usage: python <Parser>.py [ -s ] <in_xml_file>
Options:
-s Use the SAX parser, not the minidom parser.
"""
def usage():
print(USAGE_TEXT)
sys.exit(1)
def parse(inFileName):
doc = minidom.parse(inFileName)
rootNode = doc.documentElement
rootObj = DoxygenType.factory()
rootObj.build(rootNode)
# Enable Python to collect the space used by the DOM.
doc = None
sys.stdout.write('<?xml version="1.0" ?>\n')
rootObj.export(sys.stdout, 0, name_="doxygenindex", namespacedef_="")
return rootObj
def parseString(inString):
doc = minidom.parseString(inString)
rootNode = doc.documentElement
rootObj = DoxygenType.factory()
rootObj.build(rootNode)
# Enable Python to collect the space used by the DOM.
doc = None
sys.stdout.write('<?xml version="1.0" ?>\n')
rootObj.export(sys.stdout, 0, name_="doxygenindex", namespacedef_="")
return rootObj
def parseLiteral(inFileName):
doc = minidom.parse(inFileName)
rootNode = doc.documentElement
rootObj = DoxygenType.factory()
rootObj.build(rootNode)
# Enable Python to collect the space used by the DOM.
doc = None
sys.stdout.write("from index import *\n\n")
sys.stdout.write("rootObj = doxygenindex(\n")
rootObj.exportLiteral(sys.stdout, 0, name_="doxygenindex")
sys.stdout.write(")\n")
return rootObj
def main():
args = sys.argv[1:]
if len(args) == 1:
parse(args[0])
else:
usage()
if __name__ == "__main__":
main()
# import pdb
# pdb.run('main()')
|
chardet | charsetprober | ######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
# Mark Pilgrim - port to Python
# Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301 USA
######################### END LICENSE BLOCK #########################
import re
from . import constants
class CharSetProber:
def __init__(self):
pass
def reset(self):
self._mState = constants.eDetecting
def get_charset_name(self):
return None
def feed(self, aBuf):
pass
def get_state(self):
return self._mState
def get_confidence(self):
return 0.0
def filter_high_bit_only(self, aBuf):
aBuf = re.sub(b"([\x00-\x7F])+", b" ", aBuf)
return aBuf
def filter_without_english_letters(self, aBuf):
aBuf = re.sub(b"([A-Za-z])+", b" ", aBuf)
return aBuf
def filter_with_english_letters(self, aBuf):
# TODO
return aBuf
|
draftobjects | block | # ***************************************************************************
# * Copyright (c) 2009, 2010 Yorik van Havre <yorik@uncreated.net> *
# * Copyright (c) 2009, 2010 Ken Cline <cline@frii.com> *
# * Copyright (c) 2020 FreeCAD Developers *
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU Lesser General Public License (LGPL) *
# * as published by the Free Software Foundation; either version 2 of *
# * the License, or (at your option) any later version. *
# * for detail see the LICENCE text file. *
# * *
# * This program is distributed in the hope that it will be useful, *
# * but WITHOUT ANY WARRANTY; without even the implied warranty of *
# * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
# * GNU Library General Public License for more details. *
# * *
# * You should have received a copy of the GNU Library General Public *
# * License along with this program; if not, write to the Free Software *
# * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *
# * USA *
# * *
# ***************************************************************************
"""Provides the object code for the Block object."""
## @package block
# \ingroup draftobjects
# \brief Provides the object code for the Block object.
from draftobjects.base import DraftObject
## \addtogroup draftobjects
# @{
from PySide.QtCore import QT_TRANSLATE_NOOP
class Block(DraftObject):
"""The Block object"""
def __init__(self, obj):
super(Block, self).__init__(obj, "Block")
_tip = QT_TRANSLATE_NOOP("App::Property", "The components of this block")
obj.addProperty("App::PropertyLinkList", "Components", "Draft", _tip)
def execute(self, obj):
if self.props_changed_placement_only(obj):
obj.positionBySupport()
self.props_changed_clear()
return
import Part
plm = obj.Placement
shps = []
for c in obj.Components:
shps.append(c.Shape)
if shps:
shape = Part.makeCompound(shps)
obj.Shape = shape
obj.Placement = plm
obj.positionBySupport()
self.props_changed_clear()
def onChanged(self, obj, prop):
self.props_changed_store(prop)
# Alias for compatibility with v0.18 and earlier
_Block = Block
## @}
|
Material | InitGui | # ***************************************************************************
# * Copyright (c) 2013 Juergen Riegel <FreeCAD@juergen-riegel.net> *
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU Lesser General Public License (LGPL) *
# * as published by the Free Software Foundation; either version 2 of *
# * the License, or (at your option) any later version. *
# * for detail see the LICENCE text file. *
# * *
# * This program is distributed in the hope that it will be useful, *
# * but WITHOUT ANY WARRANTY; without even the implied warranty of *
# * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
# * GNU Library General Public License for more details. *
# * *
# * You should have received a copy of the GNU Library General Public *
# * License along with this program; if not, write to the Free Software *
# * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *
# * USA *
# * *
# ***************************************************************************
"""Initialization of the Material Workbench graphical interface."""
import os
import FreeCAD as App
import FreeCADGui as Gui
class MaterialWorkbench(Gui.Workbench):
"""Part workbench object."""
def __init__(self):
self.__class__.Icon = os.path.join(
App.getResourceDir(),
"Mod",
"Material",
"Resources",
"icons",
"MaterialWorkbench.svg",
)
self.__class__.MenuText = "Material"
self.__class__.ToolTip = "Material workbench"
def Initialize(self):
# load the module
import MatGui
def GetClassName(self):
return "MatGui::Workbench"
Gui.addWorkbench(MaterialWorkbench())
|
example-database-migrations | env | # -*- coding: utf-8 -*-
from __future__ import with_statement
import os
from logging.config import fileConfig
from alembic import context
from sqlalchemy import engine_from_config, pool
# this is the Alembic Config object, which provides
# access to the values within the .ini file in use.
config = context.config
# Interpret the config file for Python logging.
# This line sets up loggers basically.
fileConfig(config.config_file_name)
# add your model's MetaData object here
# for 'autogenerate' support
# from myapp import mymodel
# target_metadata = mymodel.Base.metadata
target_metadata = None
# other values from the config, defined by the needs of env.py,
# can be acquired:
# my_important_option = config.get_main_option("my_important_option")
# ... etc.
name = os.path.basename(os.path.dirname(__file__))
def run_migrations_offline():
"""Run migrations in 'offline' mode.
This configures the context with just a URL
and not an Engine, though an Engine is acceptable
here as well. By skipping the Engine creation
we don't even need a DBAPI to be available.
Calls to context.execute() here emit the given string to the
script output.
"""
url = config.get_main_option("sqlalchemy.url")
context.configure(
url=url,
target_metadata=target_metadata,
literal_binds=True,
version_table="{}_alembic_version".format(name),
)
with context.begin_transaction():
context.run_migrations()
def run_migrations_online():
"""Run migrations in 'online' mode.
In this scenario we need to create an Engine
and associate a connection with the context.
"""
connectable = engine_from_config(
config.get_section(config.config_ini_section),
prefix="sqlalchemy.",
poolclass=pool.NullPool,
)
with connectable.connect() as connection:
context.configure(
connection=connection,
target_metadata=target_metadata,
version_table="{}_alembic_version".format(name),
)
with context.begin_transaction():
context.run_migrations()
if context.is_offline_mode():
run_migrations_offline()
else:
run_migrations_online()
|
Gui | Camotics | # -*- coding: utf-8 -*-
# ***************************************************************************
# * Copyright (c) 2020 sliptonic <shopinthewoods@gmail.com> *
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU Lesser General Public License (LGPL) *
# * as published by the Free Software Foundation; either version 2 of *
# * the License, or (at your option) any later version. *
# * for detail see the LICENCE text file. *
# * *
# * This program is distributed in the hope that it will be useful, *
# * but WITHOUT ANY WARRANTY; without even the implied warranty of *
# * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
# * GNU Library General Public License for more details. *
# * *
# * You should have received a copy of the GNU Library General Public *
# * License along with this program; if not, write to the Free Software *
# * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *
# * USA *
# * *
# ***************************************************************************
import io
import json
import queue
import subprocess
from threading import Lock, Thread
import camotics
import FreeCAD
import FreeCADGui
import Mesh
import Path
import Path.Post.Command as PathPost
import PathScripts
from PySide import QtCore, QtGui
from PySide.QtCore import QT_TRANSLATE_NOOP
__title__ = "Camotics Simulator"
__author__ = "sliptonic (Brad Collette)"
__url__ = "https://www.freecad.org"
__doc__ = "Task panel for Camotics Simulation"
if False:
Path.Log.setLevel(Path.Log.Level.DEBUG, Path.Log.thisModule())
Path.Log.trackModule(Path.Log.thisModule())
else:
Path.Log.setLevel(Path.Log.Level.INFO, Path.Log.thisModule())
translate = FreeCAD.Qt.translate
class CAMoticsUI:
def __init__(self, simulation):
# this will create a Qt widget from our ui file
self.form = FreeCADGui.PySideUic.loadUi(":/panels/TaskPathCamoticsSim.ui")
self.simulation = simulation
self.initializeUI()
self.lock = False
def initializeUI(self):
self.form.timeSlider.sliderReleased.connect(
lambda: self.simulation.execute(self.form.timeSlider.value())
)
self.form.progressBar.reset()
self.form.timeSlider.setEnabled = False
self.form.btnLaunchCamotics.clicked.connect(self.launchCamotics)
self.form.btnMakeFile.clicked.connect(self.makeCamoticsFile)
self.simulation.progressUpdate.connect(self.calculating)
self.simulation.statusChange.connect(self.updateStatus)
self.form.txtStatus.setText(translate("Path", "Drag Slider to Simulate"))
def launchCamotics(self):
filename = self.makeCamoticsFile()
subprocess.Popen(["camotics", filename])
def makeCamoticsFile(self):
Path.Log.track()
filename = QtGui.QFileDialog.getSaveFileName(
self.form,
translate("Path", "Save Project As"),
"",
translate("Path", "Camotics Project (*.camotics)"),
)[0]
if filename:
if not filename.endswith(".camotics"):
filename += ".camotics"
text = self.simulation.buildproject()
try:
with open(filename, "w") as outputfile:
outputfile.write(text)
except IOError:
QtGui.QMessageBox.information(
self, translate("Path", "Unable to open file: {}".format(filename))
)
return filename
def accept(self):
self.simulation.accept()
FreeCADGui.Control.closeDialog()
def reject(self):
self.simulation.cancel()
if self.simulation.simmesh is not None:
FreeCAD.ActiveDocument.removeObject(self.simulation.simmesh.Name)
FreeCADGui.Control.closeDialog()
def setRunTime(self, duration):
self.form.timeSlider.setMinimum(0)
self.form.timeSlider.setMaximum(duration)
def calculating(self, progress=0.0):
self.form.timeSlider.setEnabled = progress == 1.0
self.form.progressBar.setValue(int(progress * 100))
def updateStatus(self, status):
self.form.txtStatus.setText(status)
class CamoticsSimulation(QtCore.QObject):
SIM = camotics.Simulation()
q = queue.Queue()
progressUpdate = QtCore.Signal(object)
statusChange = QtCore.Signal(object)
simmesh = None
filenames = []
SHAPEMAP = {
"ballend": "Ballnose",
"endmill": "Cylindrical",
"v-bit": "Conical",
"chamfer": "Snubnose",
}
def worker(self, lock):
while True:
item = self.q.get()
Path.Log.debug("worker processing: {}".format(item))
with lock:
if item["TYPE"] == "STATUS":
self.statusChange.emit(item["VALUE"])
if item["VALUE"] == "DONE":
self.SIM.wait()
surface = self.SIM.get_surface("binary")
self.SIM.wait()
self.addMesh(surface)
elif item["TYPE"] == "PROGRESS":
self.progressUpdate.emit(item["VALUE"])
self.q.task_done()
def __init__(self):
super().__init__() # needed for QT signals
lock = Lock()
Thread(target=self.worker, daemon=True, args=(lock,)).start()
def callback(self, status, progress):
self.q.put({"TYPE": "PROGRESS", "VALUE": progress})
self.q.put({"TYPE": "STATUS", "VALUE": status})
def isDone(self, success):
self.q.put({"TYPE": "STATUS", "VALUE": "DONE"})
def addMesh(self, surface):
"""takes a binary stl and adds a Mesh to the current document"""
if self.simmesh is None:
self.simmesh = FreeCAD.ActiveDocument.addObject("Mesh::Feature", "Camotics")
buffer = io.BytesIO()
buffer.write(surface)
buffer.seek(0)
mesh = Mesh.Mesh()
mesh.read(buffer, "STL")
self.simmesh.Mesh = mesh
# Mesh.show(mesh)
def Activate(self):
self.taskForm = CAMoticsUI(self)
FreeCADGui.Control.showDialog(self.taskForm)
self.job = FreeCADGui.Selection.getSelectionEx()[0].Object
self.SIM.set_metric()
self.SIM.set_resolution("high")
bb = self.job.Stock.Shape.BoundBox
self.SIM.set_workpiece(
min=(bb.XMin, bb.YMin, bb.ZMin), max=(bb.XMax, bb.YMax, bb.ZMax)
)
for t in self.job.Tools.Group:
self.SIM.set_tool(
t.ToolNumber,
metric=True,
shape=self.SHAPEMAP.get(t.Tool.ShapeName, "Cylindrical"),
length=t.Tool.Length.Value,
diameter=t.Tool.Diameter.Value,
)
postlist = PathPost.buildPostList(self.job)
Path.Log.track(postlist)
# self.filenames = [PathPost.resolveFileName(self.job)]
success = True
finalgcode = ""
for idx, section in enumerate(postlist):
partname = section[0]
sublist = section[1]
result, gcode, name = PathPost.CommandPathPost().exportObjectsWith(
sublist,
partname,
self.job,
idx,
extraargs="--no-show-editor",
)
self.filenames.append(name)
Path.Log.track(result, gcode, name)
if result is None:
success = False
else:
finalgcode += gcode
if not success:
return
self.SIM.compute_path(finalgcode)
self.SIM.wait()
tot = sum([step["time"] for step in self.SIM.get_path()])
Path.Log.debug("sim time: {}".format(tot))
self.taskForm.setRunTime(tot)
def execute(self, timeIndex):
Path.Log.track()
self.SIM.start(self.callback, time=timeIndex, done=self.isDone)
def accept(self):
pass
def cancel(self):
pass
def buildproject(self): # , files=[]):
Path.Log.track()
job = self.job
tooltemplate = {
"units": "metric",
"shape": "cylindrical",
"length": 10,
"diameter": 3.125,
"description": "",
}
workpiecetemplate = {
"automatic": False,
"margin": 0,
"bounds": {"min": [0, 0, 0], "max": [0, 0, 0]},
}
camoticstemplate = {
"units": "metric",
"resolution-mode": "medium",
"resolution": 1,
"tools": {},
"workpiece": {},
"files": [],
}
unitstring = (
"imperial" if FreeCAD.Units.getSchema() in [2, 3, 5, 7] else "metric"
)
camoticstemplate["units"] = unitstring
camoticstemplate["resolution-mode"] = "medium"
camoticstemplate["resolution"] = 1
toollist = {}
for t in job.Tools.Group:
toolitem = tooltemplate.copy()
toolitem["units"] = unitstring
if hasattr(t.Tool, "Camotics"):
toolitem["shape"] = t.Tool.Camotics
else:
toolitem["shape"] = self.SHAPEMAP.get(t.Tool.ShapeName, "Cylindrical")
toolitem["length"] = t.Tool.Length.Value
toolitem["diameter"] = t.Tool.Diameter.Value
toolitem["description"] = t.Label
toollist[t.ToolNumber] = toolitem
camoticstemplate["tools"] = toollist
bb = job.Stock.Shape.BoundBox
workpiecetemplate["bounds"]["min"] = [bb.XMin, bb.YMin, bb.ZMin]
workpiecetemplate["bounds"]["max"] = [bb.XMax, bb.YMax, bb.ZMax]
camoticstemplate["workpiece"] = workpiecetemplate
camoticstemplate["files"] = self.filenames # files
return json.dumps(camoticstemplate, indent=2)
class CommandCamoticsSimulate:
def GetResources(self):
return {
"Pixmap": "Path_Camotics",
"MenuText": QT_TRANSLATE_NOOP("Path_Camotics", "Camotics"),
"Accel": "P, C",
"ToolTip": QT_TRANSLATE_NOOP("Path_Camotics", "Simulate using Camotics"),
"CmdType": "ForEdit",
}
def IsActive(self):
if bool(FreeCADGui.Selection.getSelection()) is False:
return False
try:
job = FreeCADGui.Selection.getSelectionEx()[0].Object
return isinstance(job.Proxy, Path.Main.Job.ObjectJob)
except:
return False
def Activated(self):
pathSimulation = CamoticsSimulation()
pathSimulation.Activate()
if FreeCAD.GuiUp:
FreeCADGui.addCommand("Path_Camotics", CommandCamoticsSimulate())
FreeCAD.Console.PrintLog("Loading PathCamoticsSimulateGui ... done\n")
|
services | authentication | # -*- coding: utf-8 -*-
"""
flaskbb.auth.services.authentication
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Authentication providers, handlers and post-processors
in FlaskBB
:copyright: (c) 2014-2018 the FlaskBB Team.
:license: BSD, see LICENSE for more details
"""
import logging
from datetime import datetime
import attr
from flask_babelplus import gettext as _
from pytz import UTC
from werkzeug.security import check_password_hash
from ...core.auth.authentication import (
AuthenticationFailureHandler,
AuthenticationManager,
AuthenticationProvider,
PostAuthenticationHandler,
StopAuthentication,
)
from ...extensions import db
from ...user.models import User
from ...utils.helpers import time_utcnow
logger = logging.getLogger(__name__)
@attr.s(frozen=True)
class FailedLoginConfiguration(object):
"""
Used to configure how many failed logins are accepted until an account
is temporarily locked out and how long to temporarily lock the account
out for.
"""
limit = attr.ib()
lockout_window = attr.ib()
class BlockTooManyFailedLogins(AuthenticationProvider):
"""
Pre authentication check to block a login from an account that has too
many failed login attempts in place.
"""
def __init__(self, configuration):
self.configuration = configuration
def authenticate(self, identifier, secret):
user = User.query.filter(
db.or_(User.username == identifier, User.email == identifier)
).first()
if user is not None:
attempts = user.login_attempts
last_attempt = user.last_failed_login or datetime.min.replace(tzinfo=UTC)
reached_attempt_limit = attempts >= self.configuration.limit
inside_lockout = (
last_attempt + self.configuration.lockout_window
) >= time_utcnow()
if reached_attempt_limit and inside_lockout:
raise StopAuthentication(
_(
"Your account is currently locked out due to too many "
"failed login attempts"
)
)
class DefaultFlaskBBAuthProvider(AuthenticationProvider):
"""
This is the default username/email and password authentication checker,
locates the user based on the identifer passed -- either username or email
-- and compares the supplied password to the hash connected to the matching
user (if any).
Offers protection against timing attacks that would rely on the difference
in response time from not matching a password hash.
"""
def authenticate(self, identifier, secret):
user = User.query.filter(
db.or_(User.username == identifier, User.email == identifier)
).first()
if user is not None:
if check_password_hash(user.password, secret):
return user
return None
check_password_hash("dummy password", secret)
return None
class MarkFailedLogin(AuthenticationFailureHandler):
"""
Failure handler that marks the login attempt on the user and sets the
last failed date when it happened.
"""
def handle_authentication_failure(self, identifier):
user = User.query.filter(
db.or_(User.username == identifier, User.email == identifier)
).first()
if user is not None:
user.login_attempts += 1
user.last_failed_login = time_utcnow()
class BlockUnactivatedUser(PostAuthenticationHandler):
"""
Post auth handler that will block a user that has managed to pass the
authentication check but has not actually activated their account yet.
"""
def handle_post_auth(self, user):
if not user.activated: # pragma: no branch
raise StopAuthentication(
_(
"In order to use your account you have to "
"activate it through the link we have sent to "
"your email address."
)
)
class ClearFailedLogins(PostAuthenticationHandler):
"""
Post auth handler that clears all failed login attempts from a user's
account.
"""
def handle_post_auth(self, user):
user.login_attempts = 0
class PluginAuthenticationManager(AuthenticationManager):
"""
Authentication manager relying on plugin hooks to manage the authentication
process. This is the default authentication manager for FlaskBB.
"""
def __init__(self, plugin_manager, session):
self.plugin_manager = plugin_manager
self.session = session
def authenticate(self, identifier, secret):
try:
user = self.plugin_manager.hook.flaskbb_authenticate(
identifier=identifier, secret=secret
)
if user is None:
raise StopAuthentication(_("Wrong username or password."))
self.plugin_manager.hook.flaskbb_post_authenticate(user=user)
return user
except StopAuthentication:
self.plugin_manager.hook.flaskbb_authentication_failed(
identifier=identifier
)
raise
finally:
try:
self.session.commit()
except Exception:
logger.exception("Exception while processing login")
self.session.rollback()
raise
|
beetsplug | fuzzy | # This file is part of beets.
# Copyright 2016, Philippe Mongeau.
#
# Permission is hereby granted, free of charge, to any person obtaining
# a copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish,
# distribute, sublicense, and/or sell copies of the Software, and to
# permit persons to whom the Software is furnished to do so, subject to
# the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
"""Provides a fuzzy matching query.
"""
import difflib
from beets import config
from beets.dbcore.query import StringFieldQuery
from beets.plugins import BeetsPlugin
class FuzzyQuery(StringFieldQuery):
@classmethod
def string_match(cls, pattern, val):
# smartcase
if pattern.islower():
val = val.lower()
query_matcher = difflib.SequenceMatcher(None, pattern, val)
threshold = config["fuzzy"]["threshold"].as_number()
return query_matcher.quick_ratio() >= threshold
class FuzzyPlugin(BeetsPlugin):
def __init__(self):
super().__init__()
self.config.add(
{
"prefix": "~",
"threshold": 0.7,
}
)
def queries(self):
prefix = self.config["prefix"].as_str()
return {prefix: FuzzyQuery}
|
Extractor | setup | #
# Copyright (C) 2009 Andrew Resch <andrewresch@gmail.com>
#
# Basic plugin template created by:
# Copyright (C) 2008 Martijn Voncken <mvoncken@gmail.com>
# Copyright (C) 2007-2009 Andrew Resch <andrewresch@gmail.com>
#
# This file is part of Deluge and is licensed under GNU General Public License 3.0, or later, with
# the additional special exception to link portions of this program with the OpenSSL library.
# See LICENSE for more details.
#
from setuptools import find_packages, setup
__plugin_name__ = "Extractor"
__author__ = "Andrew Resch"
__author_email__ = "andrewresch@gmail.com"
__version__ = "0.7"
__url__ = "http://deluge-torrent.org"
__license__ = "GPLv3"
__description__ = "Extract files upon torrent completion"
__long_description__ = """
Extract files upon torrent completion
Supports: .rar, .tar, .zip, .7z .tar.gz, .tgz, .tar.bz2, .tbz .tar.lzma, .tlz, .tar.xz, .txz
Windows support: .rar, .zip, .tar, .7z, .xz, .lzma
( Requires 7-zip installed: http://www.7-zip.org/ )
Note: Will not extract with 'Move Completed' enabled
"""
__pkg_data__ = {"deluge_" + __plugin_name__.lower(): ["data/*"]}
setup(
name=__plugin_name__,
version=__version__,
description=__description__,
author=__author__,
author_email=__author_email__,
url=__url__,
license=__license__,
long_description=__long_description__ if __long_description__ else __description__,
packages=find_packages(),
package_data=__pkg_data__,
entry_points="""
[deluge.plugin.core]
%s = deluge_%s:CorePlugin
[deluge.plugin.gtk3ui]
%s = deluge_%s:GtkUIPlugin
[deluge.plugin.web]
%s = deluge_%s:WebUIPlugin
"""
% ((__plugin_name__, __plugin_name__.lower()) * 3),
)
|
forum | views | """Forum views."""
from __future__ import annotations
from collections import Counter
from datetime import date, datetime, timedelta
from itertools import groupby
from urllib.parse import quote
import sqlalchemy as sa
from abilian.core.util import utc_dt
from abilian.i18n import _, _l
from abilian.sbe.apps.communities.security import is_manager
from abilian.services.viewtracker import viewtracker
from abilian.web import url_for, views
from abilian.web.action import ButtonAction, Endpoint
from abilian.web.nav import BreadcrumbItem
from abilian.web.views import default_view
from flask import current_app, flash, g, make_response, render_template, request
from flask_babel import format_date
from flask_login import current_user
from sqlalchemy.orm import joinedload
from werkzeug.exceptions import BadRequest, NotFound
from ..communities.blueprint import Blueprint
from ..communities.common import activity_time_format, object_viewers
from ..communities.views import default_view_kw
from .forms import PostEditForm, PostForm, ThreadForm
from .models import Post, PostAttachment, Thread
from .tasks import send_post_by_email
# TODO: move to config
MAX_THREADS = 30
forum = Blueprint("forum", __name__, url_prefix="/forum", template_folder="templates")
route = forum.route
def post_kw_view_func(kw, obj, obj_type, obj_id, **kwargs):
"""kwargs for Post default view."""
kw = default_view_kw(kw, obj.thread, obj_type, obj_id, **kwargs)
kw["thread_id"] = obj.thread_id
kw["_anchor"] = f"post_{obj.id:d}"
return kw
@forum.url_value_preprocessor
def init_forum_values(endpoint, values):
g.current_tab = "forum"
g.breadcrumb.append(
BreadcrumbItem(
label=_l("Conversations"),
url=Endpoint("forum.index", community_id=g.community.slug),
)
)
def get_nb_viewers(entities):
if entities:
views = viewtracker.get_views(entities=entities)
threads = [
thread.entity
for thread in views
if thread.user in g.community.members
and thread.user != thread.entity.creator
]
return Counter(threads)
def get_viewed_posts(entities):
if not entities:
return
views = viewtracker.get_views(entities=entities, user=current_user)
all_hits = viewtracker.get_hits(views=views)
nb_viewed_posts = {}
for view in views:
related_hits = [hit for hit in all_hits if hit.view_id == view.id]
entity = view.entity
if entity in entities:
cutoff = related_hits[-1].viewed_at
nb_viewed_posts[entity] = len(
[post for post in entity.posts if post.created_at > cutoff]
)
never_viewed = set(entities) - {view.entity for view in views}
for entity in never_viewed:
nb_viewed_posts[entity] = len(entity.posts) - 1
return nb_viewed_posts
def get_viewed_times(entities):
if entities:
views = viewtracker.get_views(entities=entities)
views = [
view
for view in views
if view.user != view.entity.creator and view.user in g.community.members
]
all_hits = viewtracker.get_hits(views=views)
views_id = [view.view_id for view in all_hits]
viewed_times = Counter(views_id)
entity_viewed_times = {}
for view in views:
if view.entity not in entity_viewed_times:
entity_viewed_times[view.entity] = viewed_times[view.id]
else:
entity_viewed_times[view.entity] += viewed_times[view.id]
return entity_viewed_times
@route("/")
@route("/<string:filter>")
def index(filter=None):
query = Thread.query.filter(Thread.community_id == g.community.id).order_by(
Thread.last_post_at.desc()
)
threads = query.all()
has_more = False
nb_viewed_times = get_viewed_times(threads)
for thread in threads:
thread.nb_views = nb_viewed_times.get(thread, 0)
dt = None
if filter == "today":
dt = timedelta(days=1)
elif filter == "week":
dt = timedelta(days=7)
elif filter == "month":
dt = timedelta(days=31)
elif filter == "year":
dt = timedelta(days=365)
elif filter:
raise BadRequest()
if dt:
cutoff_date = datetime.utcnow() - dt
threads = [thread for thread in threads if thread.created_at > cutoff_date]
if dt:
threads = sorted(threads, key=lambda thread: -thread.nb_views)
else:
has_more = query.count() > MAX_THREADS
threads = query.limit(MAX_THREADS).all()
nb_viewers = get_nb_viewers(threads)
nb_viewed_posts = get_viewed_posts(threads)
return render_template(
"forum/index.html",
threads=threads,
has_more=has_more,
nb_viewers=nb_viewers,
nb_viewed_posts=nb_viewed_posts,
nb_viewed_times=nb_viewed_times,
activity_time_format=activity_time_format,
)
def group_monthly(entities_list):
# We're using Python's groupby instead of SA's group_by here
# because it's easier to support both SQLite and Postgres this way.
def grouper(entity):
return entity.created_at.year, entity.created_at.month
def format_month(year, month):
month = format_date(date(year, month, 1), "MMMM").capitalize()
return f"{month} {year}"
grouped_entities = groupby(entities_list, grouper)
grouped_entities = [
(format_month(year, month), list(entities))
for (year, month), entities in grouped_entities
]
return grouped_entities
@route("/archives/")
def archives():
all_threads = (
Thread.query.filter(Thread.community_id == g.community.id)
.order_by(Thread.created_at.desc())
.all()
)
grouped_threads = group_monthly(all_threads)
return render_template("forum/archives.html", grouped_threads=grouped_threads)
@route("/attachments/")
def attachments():
all_threads = (
Thread.query.filter(Thread.community_id == g.community.id)
.options(joinedload("posts"))
.options(joinedload("posts.attachments"))
.order_by(Thread.created_at.desc())
.all()
)
posts_with_attachments = []
for thread in all_threads:
for post in thread.posts:
if getattr(post, "attachments", None):
posts_with_attachments.append(post)
posts_with_attachments.sort(key=lambda post: post.created_at)
posts_with_attachments.reverse()
grouped_posts = group_monthly(posts_with_attachments)
return render_template("forum/attachments.html", grouped_posts=grouped_posts)
class BaseThreadView:
Model = Thread
Form = ThreadForm
pk = "thread_id"
base_template = "community/_base.html"
def can_send_by_mail(self):
return g.community.type == "participative" or is_manager(user=current_user)
def prepare_args(self, args, kwargs):
args, kwargs = super().prepare_args(args, kwargs)
self.send_by_email = False
if not self.can_send_by_mail() and "send_by_email" in self.form:
# remove from html form and avoid validation errors
del self.form["send_by_email"]
return args, kwargs
def index_url(self):
return url_for(".index", community_id=g.community.slug)
def view_url(self):
return url_for(self.obj)
class ThreadView(BaseThreadView, views.ObjectView):
methods = ["GET", "HEAD"]
Form = PostForm
template = "forum/thread.html"
@property
def template_kwargs(self):
kw = super().template_kwargs
kw["thread"] = self.obj
kw["is_closed"] = self.obj.closed
kw["is_manager"] = is_manager(user=current_user)
kw["viewers"] = object_viewers(self.obj)
kw["views"] = get_viewed_times([self.obj])
kw["participants"] = {post.creator for post in self.obj.posts}
kw["activity_time_format"] = activity_time_format
viewtracker.record_hit(entity=self.obj, user=current_user)
return kw
thread_view = ThreadView.as_view("thread")
default_view(forum, Post, None, kw_func=post_kw_view_func)(thread_view)
default_view(forum, Thread, "thread_id", kw_func=default_view_kw)(thread_view)
route("/<int:thread_id>/")(thread_view)
route("/<int:thread_id>/attachments")(
ThreadView.as_view("thread_attachments", template="forum/thread_attachments.html")
)
class ThreadCreate(BaseThreadView, views.ObjectCreate):
base_template = "community/_forumbase.html"
template = "forum/thread_create.html"
POST_BUTTON = ButtonAction(
"form", "create", btn_class="primary", title=_l("Post this message")
)
title = _("New conversation")
def init_object(self, args, kwargs):
args, kwargs = super().init_object(args, kwargs)
self.thread = self.obj
return args, kwargs
def before_populate_obj(self):
del self.form["attachments"]
self.message_body = self.form.message.data
del self.form["message"]
if "send_by_email" in self.form:
self.send_by_email = (
self.can_send_by_mail() and self.form.send_by_email.data
)
del self.form["send_by_email"]
def after_populate_obj(self):
if self.thread.community is None:
self.thread.community = g.community._model
self.post = self.thread.create_post(body_html=self.message_body)
obj_meta = self.post.meta.setdefault("abilian.sbe.forum", {})
obj_meta["origin"] = "web"
obj_meta["send_by_email"] = self.send_by_email
session = sa.orm.object_session(self.thread)
uploads = current_app.extensions["uploads"]
for handle in request.form.getlist("attachments"):
fileobj = uploads.get_file(current_user, handle)
if fileobj is None:
continue
meta = uploads.get_metadata(current_user, handle)
name = meta.get("filename", handle)
mimetype = meta.get("mimetype")
if not isinstance(name, str):
name = str(name, encoding="utf-8", errors="ignore")
if not name:
continue
attachment = PostAttachment(name=name)
attachment.post = self.post
with fileobj.open("rb") as f:
attachment.set_content(f.read(), mimetype)
session.add(attachment)
def commit_success(self):
if self.send_by_email:
task = send_post_by_email.delay(self.post.id)
meta = self.post.meta.setdefault("abilian.sbe.forum", {})
meta["send_post_by_email_task"] = task.id
self.post.meta.changed()
session = sa.orm.object_session(self.post)
session.commit()
@property
def activity_target(self):
return self.thread.community
def get_form_buttons(self, *args, **kwargs):
return [self.POST_BUTTON, views.object.CANCEL_BUTTON]
route("/new_thread/")(ThreadCreate.as_view("new_thread", view_endpoint=".thread"))
class ThreadPostCreate(ThreadCreate):
"""Add a new post to a thread."""
methods = ["POST"]
Form = PostForm
Model = Post
def init_object(self, args, kwargs):
# we DO want to skip ThreadCreate.init_object. hence super is not based on
# ThreadPostCreate
args, kwargs = super().init_object(args, kwargs)
thread_id = kwargs.pop(self.pk, None)
self.thread = Thread.query.get(thread_id)
Thread.query.filter(Thread.id == thread_id).update(
{Thread.last_post_at: datetime.utcnow()}
)
return args, kwargs
def after_populate_obj(self):
super().after_populate_obj()
session = sa.orm.object_session(self.obj)
session.expunge(self.obj)
self.obj = self.post
class ThreadViewers(ThreadView):
template = "forum/thread_viewers.html"
route("/<int:thread_id>/")(
ThreadPostCreate.as_view("thread_post", view_endpoint=".thread")
)
route("/<int:thread_id>/viewers")(ThreadViewers.as_view("thread_viewers"))
class ThreadDelete(BaseThreadView, views.ObjectDelete):
methods = ["POST"]
_message_success = _('Thread "{title}" deleted.')
def message_success(self):
return str(self._message_success).format(title=self.obj.title)
route("/<int:thread_id>/delete")(ThreadDelete.as_view("thread_delete"))
class ThreadCloseView(BaseThreadView, views.object.BaseObjectView):
"""Close / Re-open a thread."""
methods = ["POST"]
_VALID_ACTIONS = {"close", "reopen"}
CLOSED_MSG = _l("The thread is now closed for edition and new " "contributions.")
REOPENED_MSG = _l(
"The thread is now re-opened for edition and new " "contributions."
)
def prepare_args(self, args, kwargs):
args, kwargs = super().prepare_args(args, kwargs)
action = kwargs["action"] = request.form.get("action")
if action not in self._VALID_ACTIONS:
raise BadRequest(f"Unknown action: {action!r}")
return args, kwargs
def post(self, action=None):
is_closed = action == "close"
self.obj.closed = is_closed
sa.orm.object_session(self.obj).commit()
msg = self.CLOSED_MSG if is_closed else self.REOPENED_MSG
flash(str(msg))
return self.redirect(url_for(self.obj))
route("/<int:thread_id>/close")(ThreadCloseView.as_view("thread_close"))
class ThreadPostEdit(BaseThreadView, views.ObjectEdit):
Form = PostEditForm
Model = Post
pk = "object_id"
def can_send_by_mail(self):
# post edit: don't notify every time
return False
def init_object(self, args, kwargs):
# we DO want to skip ThreadCreate.init_object. hence super is not based on
# ThreadPostCreate
args, kwargs = super().init_object(args, kwargs)
thread_id = kwargs.pop("thread_id", None)
self.thread = self.obj.thread
assert thread_id == self.thread.id
return args, kwargs
def get_form_kwargs(self):
kwargs = super().get_form_kwargs()
kwargs["message"] = self.obj.body_html
return kwargs
def before_populate_obj(self):
self.message_body = self.form.message.data
del self.form["message"]
self.reason = self.form.reason.data
self.send_by_email = False
if "send_by_email" in self.form:
del self.form["send_by_email"]
self.attachments_to_remove = self.form["attachments"].delete_files_index
del self.form["attachments"]
def after_populate_obj(self):
session = sa.orm.object_session(self.obj)
uploads = current_app.extensions["uploads"]
self.obj.body_html = self.message_body
obj_meta = self.obj.meta.setdefault("abilian.sbe.forum", {})
history = obj_meta.setdefault("history", [])
history.append(
{
"user_id": current_user.id,
"user": str(current_user),
"date": utc_dt(datetime.utcnow()).isoformat(),
"reason": self.form.reason.data,
}
)
self.obj.meta["abilian.sbe.forum"] = obj_meta # trigger change for SA
attachments_to_remove = []
for idx in self.attachments_to_remove:
try:
idx = int(idx)
except ValueError:
continue
if idx > len(self.obj.attachments):
continue
attachments_to_remove.append(self.obj.attachments[idx])
for att in attachments_to_remove:
session.delete(att)
for handle in request.form.getlist("attachments"):
fileobj = uploads.get_file(current_user, handle)
if fileobj is None:
continue
meta = uploads.get_metadata(current_user, handle)
name = meta.get("filename", handle)
mimetype = meta.get("mimetype")
if not isinstance(name, str):
name = str(name, encoding="utf-8", errors="ignore")
if not name:
continue
attachment = PostAttachment(name=name, post=self.obj)
with fileobj.open("rb") as f:
attachment.set_content(f.read(), mimetype)
session.add(attachment)
route("/<int:thread_id>/<int:object_id>/edit")(ThreadPostEdit.as_view("post_edit"))
def attachment_kw_view_func(kw, obj, obj_type, obj_id, **kwargs):
post = obj.post
kw = default_view_kw(kw, post.thread, obj_type, obj_id, **kwargs)
kw["thread_id"] = post.thread_id
kw["post_id"] = post.id
return kw
@route("/<int:thread_id>/posts/<int:post_id>/attachment/<int:attachment_id>")
@default_view(forum, PostAttachment, "attachment_id", kw_func=attachment_kw_view_func)
def attachment_download(thread_id, post_id, attachment_id):
thread = Thread.query.get(thread_id)
post = Post.query.get(post_id)
attachment = PostAttachment.query.get(attachment_id)
if (
not (thread and post and attachment)
or post.thread is not thread
or attachment.post is not post
):
raise NotFound()
response = make_response(attachment.content)
response.headers["content-length"] = attachment.content_length
response.headers["content-type"] = attachment.content_type
filename = quote(attachment.name.encode("utf8"))
content_disposition = f'attachment;filename="{filename}"'
response.headers["content-disposition"] = content_disposition
return response
|
builtinViewColumns | graphColor | # =============================================================================
# Copyright (C) 2010 Diego Duclos
#
# This file is part of pyfa.
#
# pyfa is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# pyfa is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with pyfa. If not, see <http://www.gnu.org/licenses/>.
# =============================================================================
# noinspection PyPackageRequirements
import wx
from graphs.style import BASE_COLORS
from graphs.wrapper import SourceWrapper
from gui.viewColumn import ViewColumn
class GraphColor(ViewColumn):
name = "Graph Color"
def __init__(self, fittingView, params):
ViewColumn.__init__(self, fittingView)
self.resizable = False
self.size = 24
self.maxsize = self.size
self.mask = wx.LIST_MASK_TEXT
def getImageId(self, stuff):
if isinstance(stuff, SourceWrapper):
try:
colorData = BASE_COLORS[stuff.colorID]
except KeyError:
return -1
img = self.fittingView.imageList.GetImageIndex(colorData.iconName, "gui")
return img
return -1
def getToolTip(self, stuff):
if isinstance(stuff, SourceWrapper):
return "Change line color"
return ""
GraphColor.register()
|
Gui | Base | # -*- coding: utf-8 -*-
# ***************************************************************************
# * Copyright (c) 2017 sliptonic <shopinthewoods@gmail.com> *
# * *
# * This program is free software; you can redistribute it and/or modify *
# * it under the terms of the GNU Lesser General Public License (LGPL) *
# * as published by the Free Software Foundation; either version 2 of *
# * the License, or (at your option) any later version. *
# * for detail see the LICENCE text file. *
# * *
# * This program is distributed in the hope that it will be useful, *
# * but WITHOUT ANY WARRANTY; without even the implied warranty of *
# * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the *
# * GNU Library General Public License for more details. *
# * *
# * You should have received a copy of the GNU Library General Public *
# * License along with this program; if not, write to the Free Software *
# * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 *
# * USA *
# * *
# ***************************************************************************
import importlib
import FreeCAD
import FreeCADGui
import Path
import Path.Base.Gui.GetPoint as PathGetPoint
import Path.Base.Gui.Util as PathGuiUtil
import Path.Base.SetupSheet as PathSetupSheet
import Path.Base.Util as PathUtil
import Path.Main.Job as PathJob
import Path.Op.Base as PathOp
import Path.Op.Gui.Selection as PathSelection
import PathGui
import PathScripts.PathUtils as PathUtils
from PySide import QtCore, QtGui
from PySide.QtCore import QT_TRANSLATE_NOOP
__title__ = "Path Operation UI base classes"
__author__ = "sliptonic (Brad Collette)"
__url__ = "https://www.freecad.org"
__doc__ = "Base classes and framework for Path operation's UI"
translate = FreeCAD.Qt.translate
if False:
Path.Log.setLevel(Path.Log.Level.DEBUG, Path.Log.thisModule())
Path.Log.trackModule(Path.Log.thisModule())
else:
Path.Log.setLevel(Path.Log.Level.INFO, Path.Log.thisModule())
class ViewProvider(object):
"""
Generic view provider for path objects.
Deducts the icon name from operation name, brings up the TaskPanel
with pages corresponding to the operation's opFeatures() and forwards
property change notifications to the page controllers.
"""
def __init__(self, vobj, resources):
Path.Log.track()
self.deleteOnReject = True
self.OpIcon = ":/icons/%s.svg" % resources.pixmap
self.OpName = resources.name
self.OpPageModule = resources.opPageClass.__module__
self.OpPageClass = resources.opPageClass.__name__
# initialized later
self.vobj = vobj
self.Object = None
self.panel = None
def attach(self, vobj):
Path.Log.track()
self.vobj = vobj
self.Object = vobj.Object
self.panel = None
return
def deleteObjectsOnReject(self):
"""
deleteObjectsOnReject() ... return true if all objects should
be created if the user hits cancel. This is used during the initial
edit session, if the user does not press OK, it is assumed they've
changed their mind about creating the operation.
"""
Path.Log.track()
return hasattr(self, "deleteOnReject") and self.deleteOnReject
def setDeleteObjectsOnReject(self, state=False):
Path.Log.track()
self.deleteOnReject = state
return self.deleteOnReject
def setEdit(self, vobj=None, mode=0):
"""setEdit(vobj, mode=0) ... initiate editing of receivers model."""
Path.Log.track()
if 0 == mode:
if vobj is None:
vobj = self.vobj
page = self.getTaskPanelOpPage(vobj.Object)
page.setTitle(self.OpName)
page.setIcon(self.OpIcon)
selection = self.getSelectionFactory()
self.setupTaskPanel(
TaskPanel(vobj.Object, self.deleteObjectsOnReject(), page, selection)
)
self.deleteOnReject = False
return True
# no other editing possible
return False
def setupTaskPanel(self, panel):
"""setupTaskPanel(panel) ... internal function to start the editor."""
self.panel = panel
FreeCADGui.Control.closeDialog()
FreeCADGui.Control.showDialog(panel)
panel.setupUi()
job = self.Object.Proxy.getJob(self.Object)
if job:
job.ViewObject.Proxy.setupEditVisibility(job)
else:
Path.Log.info("did not find no job")
def clearTaskPanel(self):
"""clearTaskPanel() ... internal callback function when editing has finished."""
self.panel = None
job = self.Object.Proxy.getJob(self.Object)
if job:
job.ViewObject.Proxy.resetEditVisibility(job)
def unsetEdit(self, arg1, arg2):
if self.panel:
self.panel.reject(False)
def dumps(self):
"""dumps() ... callback before receiver is saved to a file.
Returns a dictionary with the receiver's resources as strings."""
Path.Log.track()
state = {}
state["OpName"] = self.OpName
state["OpIcon"] = self.OpIcon
state["OpPageModule"] = self.OpPageModule
state["OpPageClass"] = self.OpPageClass
return state
def loads(self, state):
"""loads(state) ... callback on restoring a saved instance, pendant to dumps()
state is the dictionary returned by dumps()."""
self.OpName = state["OpName"]
self.OpIcon = state["OpIcon"]
self.OpPageModule = state["OpPageModule"]
self.OpPageClass = state["OpPageClass"]
def getIcon(self):
"""getIcon() ... the icon used in the object tree"""
if self.Object.Active:
return self.OpIcon
else:
return ":/icons/Path_OpActive.svg"
def getTaskPanelOpPage(self, obj):
"""getTaskPanelOpPage(obj) ... use the stored information to instantiate the receiver op's page controller."""
mod = importlib.import_module(self.OpPageModule)
cls = getattr(mod, self.OpPageClass)
return cls(obj, 0)
def getSelectionFactory(self):
"""getSelectionFactory() ... return a factory function that can be used to create the selection observer."""
return PathSelection.select(self.OpName)
def updateData(self, obj, prop):
"""updateData(obj, prop) ... callback whenever a property of the receiver's model is assigned.
The callback is forwarded to the task panel - in case an editing session is ongoing.
"""
# Path.Log.track(obj.Label, prop) # Creates a lot of noise
if self.panel:
self.panel.updateData(obj, prop)
def onDelete(self, vobj, arg2=None):
PathUtil.clearExpressionEngine(vobj.Object)
return True
def setupContextMenu(self, vobj, menu):
Path.Log.track()
for action in menu.actions():
menu.removeAction(action)
action = QtGui.QAction(translate("PathOp", "Edit"), menu)
action.triggered.connect(self.setEdit)
menu.addAction(action)
class TaskPanelPage(object):
"""Base class for all task panel pages."""
# task panel interaction framework
def __init__(self, obj, features):
"""__init__(obj, features) ... framework initialisation.
Do not overwrite, implement initPage(obj) instead."""
self.obj = obj
self.job = PathUtils.findParentJob(obj)
self.form = self.getForm()
self.signalDirtyChanged = None
self.setClean()
self.setTitle("-")
self.setIcon(None)
self.features = features
self.isdirty = False
self.parent = None
self.panelTitle = "Operation"
if self._installTCUpdate():
PathJob.Notification.updateTC.connect(self.resetToolController)
def _installTCUpdate(self):
return hasattr(self.form, "toolController")
def setParent(self, parent):
"""setParent() ... used to transfer parent object link to child class.
Do not overwrite."""
self.parent = parent
def onDirtyChanged(self, callback):
"""onDirtyChanged(callback) ... set callback when dirty state changes."""
self.signalDirtyChanged = callback
def setDirty(self):
"""setDirty() ... mark receiver as dirty, causing the model to be recalculated if OK or Apply is pressed."""
self.isdirty = True
if self.signalDirtyChanged:
self.signalDirtyChanged(self)
def setClean(self):
"""setClean() ... mark receiver as clean, indicating there is no need to recalculate the model even if the user presses OK or Apply."""
self.isdirty = False
if self.signalDirtyChanged:
self.signalDirtyChanged(self)
def pageGetFields(self):
"""pageGetFields() ... internal callback.
Do not overwrite, implement getFields(obj) instead."""
self.getFields(self.obj)
self.setDirty()
def pageSetFields(self):
"""pageSetFields() ... internal callback.
Do not overwrite, implement setFields(obj) instead."""
self.setFields(self.obj)
def pageCleanup(self):
"""pageCleanup() ... internal callback.
Do not overwrite, implement cleanupPage(obj) instead."""
if self._installTCUpdate():
PathJob.Notification.updateTC.disconnect(self.resetToolController)
self.cleanupPage(self.obj)
def pageRegisterSignalHandlers(self):
"""pageRegisterSignalHandlers() .. internal callback.
Registers a callback for all signals returned by getSignalsForUpdate(obj).
Do not overwrite, implement getSignalsForUpdate(obj) and/or registerSignalHandlers(obj) instead.
"""
for signal in self.getSignalsForUpdate(self.obj):
signal.connect(self.pageGetFields)
self.registerSignalHandlers(self.obj)
def pageUpdateData(self, obj, prop):
"""pageUpdateData(obj, prop) ... internal callback.
Do not overwrite, implement updateData(obj) instead."""
self.updateData(obj, prop)
def setTitle(self, title):
"""setTitle(title) ... sets a title for the page."""
self.title = title
def getTitle(self, obj):
"""getTitle(obj) ... return title to be used for the receiver page.
The default implementation returns what was previously set with setTitle(title).
Can safely be overwritten by subclasses."""
return self.title
def setIcon(self, icon):
"""setIcon(icon) ... sets the icon for the page."""
self.icon = icon
def getIcon(self, obj):
"""getIcon(obj) ... return icon for page or None.
Can safely be overwritten by subclasses."""
return self.icon
# subclass interface
def initPage(self, obj):
"""initPage(obj) ... overwrite to customize UI for specific model.
Note that this function is invoked after all page controllers have been created.
Should be overwritten by subclasses."""
pass
def cleanupPage(self, obj):
"""cleanupPage(obj) ... overwrite to perform any cleanup tasks before page is destroyed.
Can safely be overwritten by subclasses."""
pass
def modifyStandardButtons(self, buttonBox):
"""modifyStandardButtons(buttonBox) ... overwrite if the task panel standard buttons need to be modified.
Can safely be overwritten by subclasses."""
pass
def getForm(self):
"""getForm() ... return UI form for this page.
Must be overwritten by subclasses."""
pass
def getFields(self, obj):
"""getFields(obj) ... overwrite to transfer values from UI to obj's properties.
Can safely be overwritten by subclasses."""
pass
def setFields(self, obj):
"""setFields(obj) ... overwrite to transfer obj's property values to UI.
Can safely be overwritten by subclasses."""
pass
def getSignalsForUpdate(self, obj):
"""getSignalsForUpdate(obj) ... return signals which, when triggered, cause the receiver to update the model.
See also registerSignalHandlers(obj)
Can safely be overwritten by subclasses."""
return []
def registerSignalHandlers(self, obj):
"""registerSignalHandlers(obj) ... overwrite to register custom signal handlers.
In case an update of a model is not the desired operation of a signal invocation
(see getSignalsForUpdate(obj)) this function can be used to register signal handlers
manually.
Can safely be overwritten by subclasses."""
pass
def updateData(self, obj, prop):
"""updateData(obj, prop) ... overwrite if the receiver needs to react to property changes that might not have been caused by the receiver itself.
Sometimes a model will recalculate properties based on a change of another property. In order to keep the UI up to date with such changes this
function can be used.
Please note that the callback is synchronous with the property assignment operation. Also note that the notification is invoked regardless of the
actual value of the property assignment. In other words it also fires if a property gets assigned the same value it already has.
Taking above observations into account the implementation has to take care that it doesn't overwrite modified UI values by invoking setFields(obj).
This can happen if a subclass unconditionally transfers all values in getFields(obj) to the model and just calls setFields(obj) in this callback.
In such a scenario the first property assignment will cause all changes in the UI of the other fields to be overwritten by setFields(obj).
You have been warned."""
pass
def updateSelection(self, obj, sel):
"""updateSelection(obj, sel) ...
overwrite to customize UI depending on current selection.
Can safely be overwritten by subclasses."""
pass
def selectInComboBox(self, name, combo):
"""selectInComboBox(name, combo) ...
helper function to select a specific value in a combo box."""
blocker = QtCore.QSignalBlocker(combo)
index = combo.currentIndex() # Save initial index
# Search using currentData and return if found
newindex = combo.findData(name)
if newindex >= 0:
combo.setCurrentIndex(newindex)
return
# if not found, search using current text
newindex = combo.findText(name, QtCore.Qt.MatchFixedString)
if newindex >= 0:
combo.setCurrentIndex(newindex)
return
# not found, return unchanged
combo.setCurrentIndex(index)
return
def populateCombobox(self, form, enumTups, comboBoxesPropertyMap):
"""populateCombobox(form, enumTups, comboBoxesPropertyMap) ... proxy for PathGuiUtil.populateCombobox()"""
PathGuiUtil.populateCombobox(form, enumTups, comboBoxesPropertyMap)
def resetToolController(self, job, tc):
if self.obj is not None:
self.obj.ToolController = tc
combo = self.form.toolController
self.setupToolController(self.obj, combo)
def setupToolController(self, obj, combo):
"""setupToolController(obj, combo) ...
helper function to setup obj's ToolController
in the given combo box."""
controllers = PathUtils.getToolControllers(self.obj)
labels = [c.Label for c in controllers]
combo.blockSignals(True)
combo.clear()
combo.addItems(labels)
combo.blockSignals(False)
if obj.ToolController is None:
obj.ToolController = PathUtils.findToolController(obj, obj.Proxy)
if obj.ToolController is not None:
self.selectInComboBox(obj.ToolController.Label, combo)
def updateToolController(self, obj, combo):
"""updateToolController(obj, combo) ...
helper function to update obj's ToolController property if a different
one has been selected in the combo box."""
tc = PathUtils.findToolController(obj, obj.Proxy, combo.currentText())
if obj.ToolController != tc:
obj.ToolController = tc
def setupCoolant(self, obj, combo):
"""setupCoolant(obj, combo) ...
helper function to setup obj's Coolant option."""
job = PathUtils.findParentJob(obj)
options = job.SetupSheet.CoolantModes
combo.blockSignals(True)
combo.clear()
combo.addItems(options)
combo.blockSignals(False)
if hasattr(obj, "CoolantMode"):
self.selectInComboBox(obj.CoolantMode, combo)
def updateCoolant(self, obj, combo):
"""updateCoolant(obj, combo) ...
helper function to update obj's Coolant property if a different
one has been selected in the combo box."""
option = combo.currentText()
if hasattr(obj, "CoolantMode"):
if obj.CoolantMode != option:
obj.CoolantMode = option
def updatePanelVisibility(self, panelTitle, obj):
"""updatePanelVisibility(panelTitle, obj) ...
Function to call the `updateVisibility()` GUI method of the
page whose panel title is as indicated."""
if hasattr(self, "parent"):
parent = getattr(self, "parent")
if parent and hasattr(parent, "featurePages"):
for page in parent.featurePages:
if hasattr(page, "panelTitle"):
if page.panelTitle == panelTitle and hasattr(
page, "updateVisibility"
):
page.updateVisibility()
break
class TaskPanelBaseGeometryPage(TaskPanelPage):
"""Page controller for the base geometry."""
DataObject = QtCore.Qt.ItemDataRole.UserRole
DataObjectSub = QtCore.Qt.ItemDataRole.UserRole + 1
def __init__(self, obj, features):
super(TaskPanelBaseGeometryPage, self).__init__(obj, features)
self.panelTitle = "Base Geometry"
self.OpIcon = ":/icons/Path_BaseGeometry.svg"
self.setIcon(self.OpIcon)
def getForm(self):
panel = FreeCADGui.PySideUic.loadUi(":/panels/PageBaseGeometryEdit.ui")
self.modifyPanel(panel)
return panel
def modifyPanel(self, panel):
"""modifyPanel(self, panel) ...
Helper method to modify the current form immediately after
it is loaded."""
# Determine if Job operations are available with Base Geometry
availableOps = list()
ops = self.job.Operations.Group
for op in ops:
if hasattr(op, "Base") and isinstance(op.Base, list):
if len(op.Base) > 0:
availableOps.append(op.Label)
# Load available operations into combobox
if len(availableOps) > 0:
# Populate the operations list
panel.geometryImportList.blockSignals(True)
panel.geometryImportList.clear()
availableOps.sort()
for opLbl in availableOps:
panel.geometryImportList.addItem(opLbl)
panel.geometryImportList.blockSignals(False)
else:
panel.geometryImportList.hide()
panel.geometryImportButton.hide()
def getTitle(self, obj):
return translate("PathOp", "Base Geometry")
def getFields(self, obj):
pass
def setFields(self, obj):
self.form.baseList.blockSignals(True)
self.form.baseList.clear()
for base in self.obj.Base:
for sub in base[1]:
item = QtGui.QListWidgetItem("%s.%s" % (base[0].Label, sub))
item.setData(self.DataObject, base[0])
item.setData(self.DataObjectSub, sub)
self.form.baseList.addItem(item)
self.form.baseList.blockSignals(False)
self.resizeBaseList()
def itemActivated(self):
FreeCADGui.Selection.clearSelection()
for item in self.form.baseList.selectedItems():
obj = item.data(self.DataObject)
sub = item.data(self.DataObjectSub)
if sub:
FreeCADGui.Selection.addSelection(obj, sub)
else:
FreeCADGui.Selection.addSelection(obj)
# FreeCADGui.updateGui()
def supportsVertexes(self):
return self.features & PathOp.FeatureBaseVertexes
def supportsEdges(self):
return self.features & PathOp.FeatureBaseEdges
def supportsFaces(self):
return self.features & PathOp.FeatureBaseFaces
def supportsPanels(self):
return self.features & PathOp.FeatureBasePanels
def featureName(self):
if self.supportsEdges() and self.supportsFaces():
return "features"
if self.supportsFaces():
return "faces"
if self.supportsEdges():
return "edges"
return "nothing"
def selectionSupportedAsBaseGeometry(self, selection, ignoreErrors):
if len(selection) != 1:
if not ignoreErrors:
msg = translate(
"PathOp",
"Please select %s from a single solid" % self.featureName(),
)
FreeCAD.Console.PrintError(msg + "\n")
Path.Log.debug(msg)
return False
sel = selection[0]
if sel.HasSubObjects:
if (
not self.supportsVertexes()
and selection[0].SubObjects[0].ShapeType == "Vertex"
):
if not ignoreErrors:
Path.Log.error(translate("PathOp", "Vertexes are not supported"))
return False
if (
not self.supportsEdges()
and selection[0].SubObjects[0].ShapeType == "Edge"
):
if not ignoreErrors:
Path.Log.error(translate("PathOp", "Edges are not supported"))
return False
if (
not self.supportsFaces()
and selection[0].SubObjects[0].ShapeType == "Face"
):
if not ignoreErrors:
Path.Log.error(translate("PathOp", "Faces are not supported"))
return False
else:
if not self.supportsPanels() or "Panel" not in sel.Object.Name:
if not ignoreErrors:
Path.Log.error(
translate(
"PathOp",
"Please select %s of a solid" % self.featureName(),
)
)
return False
return True
def addBaseGeometry(self, selection):
Path.Log.track(selection)
if self.selectionSupportedAsBaseGeometry(selection, False):
sel = selection[0]
for sub in sel.SubElementNames:
self.obj.Proxy.addBase(self.obj, sel.Object, sub)
return True
return False
def addBase(self):
Path.Log.track()
if self.addBaseGeometry(FreeCADGui.Selection.getSelectionEx()):
# self.obj.Proxy.execute(self.obj)
self.setFields(self.obj)
self.setDirty()
self.updatePanelVisibility("Operation", self.obj)
def deleteBase(self):
Path.Log.track()
selected = self.form.baseList.selectedItems()
for item in selected:
self.form.baseList.takeItem(self.form.baseList.row(item))
self.setDirty()
self.updateBase()
self.updatePanelVisibility("Operation", self.obj)
self.resizeBaseList()
def updateBase(self):
newlist = []
for i in range(self.form.baseList.count()):
item = self.form.baseList.item(i)
obj = item.data(self.DataObject)
sub = item.data(self.DataObjectSub)
if sub:
base = (obj, str(sub))
newlist.append(base)
Path.Log.debug("Setting new base: %s -> %s" % (self.obj.Base, newlist))
self.obj.Base = newlist
# self.obj.Proxy.execute(self.obj)
# FreeCAD.ActiveDocument.recompute()
def clearBase(self):
self.obj.Base = []
self.setDirty()
self.updatePanelVisibility("Operation", self.obj)
self.resizeBaseList()
def importBaseGeometry(self):
opLabel = str(self.form.geometryImportList.currentText())
ops = FreeCAD.ActiveDocument.getObjectsByLabel(opLabel)
if len(ops) > 1:
msg = translate("PathOp", "Multiple operations are labeled as")
msg += " {}\n".format(opLabel)
FreeCAD.Console.PrintWarning(msg)
for base, subList in ops[0].Base:
FreeCADGui.Selection.clearSelection()
FreeCADGui.Selection.addSelection(base, subList)
self.addBase()
def registerSignalHandlers(self, obj):
self.form.baseList.itemSelectionChanged.connect(self.itemActivated)
self.form.addBase.clicked.connect(self.addBase)
self.form.deleteBase.clicked.connect(self.deleteBase)
self.form.clearBase.clicked.connect(self.clearBase)
self.form.geometryImportButton.clicked.connect(self.importBaseGeometry)
def pageUpdateData(self, obj, prop):
if prop in ["Base"]:
self.setFields(obj)
def updateSelection(self, obj, sel):
if self.selectionSupportedAsBaseGeometry(sel, True):
self.form.addBase.setEnabled(True)
else:
self.form.addBase.setEnabled(False)
def resizeBaseList(self):
# Set base geometry list window to resize based on contents
# Code reference:
# https://stackoverflow.com/questions/6337589/qlistwidget-adjust-size-to-content
# ml: disabling this logic because I can't get it to work on HPD monitor.
# On my systems the values returned by the list object are also incorrect on
# creation, leading to a list object of size 15. count() always returns 0 until
# the list is actually displayed. The same is true for sizeHintForRow(0), which
# returns -1 until the widget is rendered. The widget claims to have a size of
# (100, 30), once it becomes visible the size is (535, 192).
# Leaving the framework here in case somebody figures out how to set this up
# properly.
qList = self.form.baseList
row = (qList.count() + qList.frameWidth()) * 15
# qList.setMinimumHeight(row)
Path.Log.debug(
"baseList({}, {}) {} * {}".format(
qList.size(), row, qList.count(), qList.sizeHintForRow(0)
)
)
class TaskPanelBaseLocationPage(TaskPanelPage):
"""Page controller for base locations. Uses PathGetPoint."""
DataLocation = QtCore.Qt.ItemDataRole.UserRole
def __init__(self, obj, features):
super(TaskPanelBaseLocationPage, self).__init__(obj, features)
# members initialized later
self.editRow = None
self.panelTitle = "Base Location"
def getForm(self):
self.formLoc = FreeCADGui.PySideUic.loadUi(":/panels/PageBaseLocationEdit.ui")
if QtCore.qVersion()[0] == "4":
self.formLoc.baseList.horizontalHeader().setResizeMode(
QtGui.QHeaderView.Stretch
)
else:
self.formLoc.baseList.horizontalHeader().setSectionResizeMode(
QtGui.QHeaderView.Stretch
)
self.getPoint = PathGetPoint.TaskPanel(self.formLoc.addRemoveEdit)
return self.formLoc
def modifyStandardButtons(self, buttonBox):
self.getPoint.buttonBox = buttonBox
def getTitle(self, obj):
return translate("PathOp", "Base Location")
def getFields(self, obj):
pass
def setFields(self, obj):
self.formLoc.baseList.blockSignals(True)
self.formLoc.baseList.clearContents()
self.formLoc.baseList.setRowCount(0)
for location in self.obj.Locations:
self.formLoc.baseList.insertRow(self.formLoc.baseList.rowCount())
item = QtGui.QTableWidgetItem("%.2f" % location.x)
item.setData(self.DataLocation, location.x)
self.formLoc.baseList.setItem(self.formLoc.baseList.rowCount() - 1, 0, item)
item = QtGui.QTableWidgetItem("%.2f" % location.y)
item.setData(self.DataLocation, location.y)
self.formLoc.baseList.setItem(self.formLoc.baseList.rowCount() - 1, 1, item)
self.formLoc.baseList.resizeColumnToContents(0)
self.formLoc.baseList.blockSignals(False)
self.itemActivated()
def removeLocation(self):
deletedRows = []
selected = self.formLoc.baseList.selectedItems()
for item in selected:
row = self.formLoc.baseList.row(item)
if row not in deletedRows:
deletedRows.append(row)
self.formLoc.baseList.removeRow(row)
self.updateLocations()
FreeCAD.ActiveDocument.recompute()
def updateLocations(self):
Path.Log.track()
locations = []
for i in range(self.formLoc.baseList.rowCount()):
x = self.formLoc.baseList.item(i, 0).data(self.DataLocation)
y = self.formLoc.baseList.item(i, 1).data(self.DataLocation)
location = FreeCAD.Vector(x, y, 0)
locations.append(location)
self.obj.Locations = locations
def addLocation(self):
self.getPoint.getPoint(self.addLocationAt)
def addLocationAt(self, point, obj):
if point:
locations = self.obj.Locations
locations.append(point)
self.obj.Locations = locations
FreeCAD.ActiveDocument.recompute()
def editLocation(self):
selected = self.formLoc.baseList.selectedItems()
if selected:
row = self.formLoc.baseList.row(selected[0])
self.editRow = row
x = self.formLoc.baseList.item(row, 0).data(self.DataLocation)
y = self.formLoc.baseList.item(row, 1).data(self.DataLocation)
start = FreeCAD.Vector(x, y, 0)
self.getPoint.getPoint(self.editLocationAt, start)
def editLocationAt(self, point, obj):
if point:
self.formLoc.baseList.item(self.editRow, 0).setData(
self.DataLocation, point.x
)
self.formLoc.baseList.item(self.editRow, 1).setData(
self.DataLocation, point.y
)
self.updateLocations()
FreeCAD.ActiveDocument.recompute()
def itemActivated(self):
if self.formLoc.baseList.selectedItems():
self.form.removeLocation.setEnabled(True)
self.form.editLocation.setEnabled(True)
else:
self.form.removeLocation.setEnabled(False)
self.form.editLocation.setEnabled(False)
def registerSignalHandlers(self, obj):
self.form.baseList.itemSelectionChanged.connect(self.itemActivated)
self.formLoc.addLocation.clicked.connect(self.addLocation)
self.formLoc.removeLocation.clicked.connect(self.removeLocation)
self.formLoc.editLocation.clicked.connect(self.editLocation)
def pageUpdateData(self, obj, prop):
if prop in ["Locations"]:
self.setFields(obj)
class TaskPanelHeightsPage(TaskPanelPage):
"""Page controller for heights."""
def __init__(self, obj, features):
super(TaskPanelHeightsPage, self).__init__(obj, features)
# members initialized later
self.clearanceHeight = None
self.safeHeight = None
self.panelTitle = "Heights"
self.OpIcon = ":/icons/Path_Heights.svg"
self.setIcon(self.OpIcon)
def getForm(self):
return FreeCADGui.PySideUic.loadUi(":/panels/PageHeightsEdit.ui")
def initPage(self, obj):
self.safeHeight = PathGuiUtil.QuantitySpinBox(
self.form.safeHeight, obj, "SafeHeight"
)
self.clearanceHeight = PathGuiUtil.QuantitySpinBox(
self.form.clearanceHeight, obj, "ClearanceHeight"
)
def getTitle(self, obj):
return translate("PathOp", "Heights")
def getFields(self, obj):
self.safeHeight.updateProperty()
self.clearanceHeight.updateProperty()
def setFields(self, obj):
self.safeHeight.updateSpinBox()
self.clearanceHeight.updateSpinBox()
def getSignalsForUpdate(self, obj):
signals = []
signals.append(self.form.safeHeight.editingFinished)
signals.append(self.form.clearanceHeight.editingFinished)
return signals
def pageUpdateData(self, obj, prop):
if prop in ["SafeHeight", "ClearanceHeight"]:
self.setFields(obj)
class TaskPanelDepthsPage(TaskPanelPage):
"""Page controller for depths."""
def __init__(self, obj, features):
super(TaskPanelDepthsPage, self).__init__(obj, features)
# members initialized later
self.startDepth = None
self.finalDepth = None
self.finishDepth = None
self.stepDown = None
self.panelTitle = "Depths"
self.OpIcon = ":/icons/Path_Depths.svg"
self.setIcon(self.OpIcon)
def getForm(self):
return FreeCADGui.PySideUic.loadUi(":/panels/PageDepthsEdit.ui")
def haveStartDepth(self):
return PathOp.FeatureDepths & self.features
def haveFinalDepth(self):
return (
PathOp.FeatureDepths & self.features
and not PathOp.FeatureNoFinalDepth & self.features
)
def haveFinishDepth(self):
return (
PathOp.FeatureDepths & self.features
and PathOp.FeatureFinishDepth & self.features
)
def haveStepDown(self):
return PathOp.FeatureStepDown & self.features
def initPage(self, obj):
if self.haveStartDepth():
self.startDepth = PathGuiUtil.QuantitySpinBox(
self.form.startDepth, obj, "StartDepth"
)
else:
self.form.startDepth.hide()
self.form.startDepthLabel.hide()
self.form.startDepthSet.hide()
if self.haveFinalDepth():
self.finalDepth = PathGuiUtil.QuantitySpinBox(
self.form.finalDepth, obj, "FinalDepth"
)
else:
if self.haveStartDepth():
self.form.finalDepth.setEnabled(False)
self.form.finalDepth.setToolTip(
translate(
"PathOp",
"FinalDepth cannot be modified for this operation.\nIf it is necessary to set the FinalDepth manually please select a different operation.",
)
)
else:
self.form.finalDepth.hide()
self.form.finalDepthLabel.hide()
self.form.finalDepthSet.hide()
if self.haveStepDown():
self.stepDown = PathGuiUtil.QuantitySpinBox(
self.form.stepDown, obj, "StepDown"
)
else:
self.form.stepDown.hide()
self.form.stepDownLabel.hide()
if self.haveFinishDepth():
self.finishDepth = PathGuiUtil.QuantitySpinBox(
self.form.finishDepth, obj, "FinishDepth"
)
else:
self.form.finishDepth.hide()
self.form.finishDepthLabel.hide()
def getTitle(self, obj):
return translate("PathOp", "Depths")
def getFields(self, obj):
if self.haveStartDepth():
self.startDepth.updateProperty()
if self.haveFinalDepth():
self.finalDepth.updateProperty()
if self.haveStepDown():
self.stepDown.updateProperty()
if self.haveFinishDepth():
self.finishDepth.updateProperty()
def setFields(self, obj):
if self.haveStartDepth():
self.startDepth.updateSpinBox()
if self.haveFinalDepth():
self.finalDepth.updateSpinBox()
if self.haveStepDown():
self.stepDown.updateSpinBox()
if self.haveFinishDepth():
self.finishDepth.updateSpinBox()
self.updateSelection(obj, FreeCADGui.Selection.getSelectionEx())
def getSignalsForUpdate(self, obj):
signals = []
if self.haveStartDepth():
signals.append(self.form.startDepth.editingFinished)
if self.haveFinalDepth():
signals.append(self.form.finalDepth.editingFinished)
if self.haveStepDown():
signals.append(self.form.stepDown.editingFinished)
if self.haveFinishDepth():
signals.append(self.form.finishDepth.editingFinished)
return signals
def registerSignalHandlers(self, obj):
if self.haveStartDepth():
self.form.startDepthSet.clicked.connect(
lambda: self.depthSet(obj, self.startDepth, "StartDepth")
)
if self.haveFinalDepth():
self.form.finalDepthSet.clicked.connect(
lambda: self.depthSet(obj, self.finalDepth, "FinalDepth")
)
def pageUpdateData(self, obj, prop):
if prop in ["StartDepth", "FinalDepth", "StepDown", "FinishDepth"]:
self.setFields(obj)
def depthSet(self, obj, spinbox, prop):
z = self.selectionZLevel(FreeCADGui.Selection.getSelectionEx())
if z is not None:
Path.Log.debug("depthSet(%s, %s, %.2f)" % (obj.Label, prop, z))
if spinbox.expression():
obj.setExpression(prop, None)
self.setDirty()
spinbox.updateSpinBox(FreeCAD.Units.Quantity(z, FreeCAD.Units.Length))
if spinbox.updateProperty():
self.setDirty()
else:
Path.Log.info("depthSet(-)")
def selectionZLevel(self, sel):
if len(sel) == 1 and len(sel[0].SubObjects) == 1:
sub = sel[0].SubObjects[0]
if "Vertex" == sub.ShapeType:
return sub.Z
if Path.Geom.isHorizontal(sub):
if "Edge" == sub.ShapeType:
return sub.Vertexes[0].Z
if "Face" == sub.ShapeType:
return sub.BoundBox.ZMax
return None
def updateSelection(self, obj, sel):
if self.selectionZLevel(sel) is not None:
self.form.startDepthSet.setEnabled(True)
self.form.finalDepthSet.setEnabled(True)
else:
self.form.startDepthSet.setEnabled(False)
self.form.finalDepthSet.setEnabled(False)
class TaskPanelDiametersPage(TaskPanelPage):
"""Page controller for diameters."""
def __init__(self, obj, features):
super(TaskPanelDiametersPage, self).__init__(obj, features)
# members initialized later
self.clearanceHeight = None
self.safeHeight = None
def getForm(self):
return FreeCADGui.PySideUic.loadUi(":/panels/PageDiametersEdit.ui")
def initPage(self, obj):
self.minDiameter = PathGuiUtil.QuantitySpinBox(
self.form.minDiameter, obj, "MinDiameter"
)
self.maxDiameter = PathGuiUtil.QuantitySpinBox(
self.form.maxDiameter, obj, "MaxDiameter"
)
def getTitle(self, obj):
return translate("PathOp", "Diameters")
def getFields(self, obj):
self.minDiameter.updateProperty()
self.maxDiameter.updateProperty()
def setFields(self, obj):
self.minDiameter.updateSpinBox()
self.maxDiameter.updateSpinBox()
def getSignalsForUpdate(self, obj):
signals = []
signals.append(self.form.minDiameter.editingFinished)
signals.append(self.form.maxDiameter.editingFinished)
return signals
def pageUpdateData(self, obj, prop):
if prop in ["MinDiameter", "MaxDiameter"]:
self.setFields(obj)
class TaskPanel(object):
"""
Generic TaskPanel implementation handling the standard Path operation layout.
This class only implements the framework and takes care of bringing all pages up and down in a controller fashion.
It implements the standard editor behaviour for OK, Cancel and Apply and tracks if the model is still in sync with
the UI.
However, all display and processing of fields is handled by the page controllers which are managed in a list. All
event callbacks and framework actions are forwarded to the page controllers in turn and each page controller is
expected to process all events concerning the data it manages.
"""
def __init__(self, obj, deleteOnReject, opPage, selectionFactory):
Path.Log.track(obj.Label, deleteOnReject, opPage, selectionFactory)
FreeCAD.ActiveDocument.openTransaction(translate("PathOp", "AreaOp Operation"))
self.obj = obj
self.deleteOnReject = deleteOnReject
self.featurePages = []
self.parent = None
# members initialized later
self.clearanceHeight = None
self.safeHeight = None
self.startDepth = None
self.finishDepth = None
self.finalDepth = None
self.stepDown = None
self.buttonBox = None
self.minDiameter = None
self.maxDiameter = None
features = obj.Proxy.opFeatures(obj)
opPage.features = features
if PathOp.FeatureBaseGeometry & features:
if hasattr(opPage, "taskPanelBaseGeometryPage"):
self.featurePages.append(
opPage.taskPanelBaseGeometryPage(obj, features)
)
else:
self.featurePages.append(TaskPanelBaseGeometryPage(obj, features))
if PathOp.FeatureLocations & features:
if hasattr(opPage, "taskPanelBaseLocationPage"):
self.featurePages.append(
opPage.taskPanelBaseLocationPage(obj, features)
)
else:
self.featurePages.append(TaskPanelBaseLocationPage(obj, features))
if PathOp.FeatureDepths & features or PathOp.FeatureStepDown & features:
if hasattr(opPage, "taskPanelDepthsPage"):
self.featurePages.append(opPage.taskPanelDepthsPage(obj, features))
else:
self.featurePages.append(TaskPanelDepthsPage(obj, features))
if PathOp.FeatureHeights & features:
if hasattr(opPage, "taskPanelHeightsPage"):
self.featurePages.append(opPage.taskPanelHeightsPage(obj, features))
else:
self.featurePages.append(TaskPanelHeightsPage(obj, features))
if PathOp.FeatureDiameters & features:
if hasattr(opPage, "taskPanelDiametersPage"):
self.featurePages.append(opPage.taskPanelDiametersPage(obj, features))
else:
self.featurePages.append(TaskPanelDiametersPage(obj, features))
self.featurePages.append(opPage)
for page in self.featurePages:
page.parent = self # save pointer to this current class as "parent"
page.initPage(obj)
page.onDirtyChanged(self.pageDirtyChanged)
taskPanelLayout = Path.Preferences.defaultTaskPanelLayout()
if taskPanelLayout < 2:
opTitle = opPage.getTitle(obj)
opPage.setTitle(translate("PathOp", "Operation"))
toolbox = QtGui.QToolBox()
if taskPanelLayout == 0:
for page in self.featurePages:
toolbox.addItem(page.form, page.getTitle(obj))
itemIdx = toolbox.count() - 1
if page.icon:
toolbox.setItemIcon(itemIdx, QtGui.QIcon(page.icon))
toolbox.setCurrentIndex(len(self.featurePages) - 1)
else:
for page in reversed(self.featurePages):
toolbox.addItem(page.form, page.getTitle(obj))
itemIdx = toolbox.count() - 1
if page.icon:
toolbox.setItemIcon(itemIdx, QtGui.QIcon(page.icon))
toolbox.setWindowTitle(opTitle)
if opPage.getIcon(obj):
toolbox.setWindowIcon(QtGui.QIcon(opPage.getIcon(obj)))
self.form = toolbox
elif taskPanelLayout == 2:
forms = []
for page in self.featurePages:
page.form.setWindowTitle(page.getTitle(obj))
forms.append(page.form)
self.form = forms
elif taskPanelLayout == 3:
forms = []
for page in reversed(self.featurePages):
page.form.setWindowTitle(page.getTitle(obj))
forms.append(page.form)
self.form = forms
self.selectionFactory = selectionFactory
self.obj = obj
self.isdirty = deleteOnReject
self.visibility = obj.ViewObject.Visibility
obj.ViewObject.Visibility = True
def isDirty(self):
"""isDirty() ... returns true if the model is not in sync with the UI anymore."""
for page in self.featurePages:
if page.isdirty:
return True
return self.isdirty
def setClean(self):
"""setClean() ... set the receiver and all its pages clean."""
self.isdirty = False
for page in self.featurePages:
page.setClean()
def accept(self, resetEdit=True):
"""accept() ... callback invoked when user presses the task panel OK button."""
self.preCleanup()
if self.isDirty():
self.panelGetFields()
FreeCAD.ActiveDocument.commitTransaction()
self.cleanup(resetEdit)
def reject(self, resetEdit=True):
"""reject() ... callback invoked when user presses the task panel Cancel button."""
self.preCleanup()
FreeCAD.ActiveDocument.abortTransaction()
if self.deleteOnReject:
FreeCAD.ActiveDocument.openTransaction(
translate("PathOp", "Uncreate AreaOp Operation")
)
try:
PathUtil.clearExpressionEngine(self.obj)
FreeCAD.ActiveDocument.removeObject(self.obj.Name)
except Exception as ee:
Path.Log.debug("{}\n".format(ee))
FreeCAD.ActiveDocument.commitTransaction()
self.cleanup(resetEdit)
return True
def preCleanup(self):
for page in self.featurePages:
page.onDirtyChanged(None)
PathSelection.clear()
FreeCADGui.Selection.removeObserver(self)
self.obj.ViewObject.Proxy.clearTaskPanel()
self.obj.ViewObject.Visibility = self.visibility
def cleanup(self, resetEdit):
"""cleanup() ... implements common cleanup tasks."""
self.panelCleanup()
FreeCADGui.Control.closeDialog()
if resetEdit:
FreeCADGui.ActiveDocument.resetEdit()
FreeCAD.ActiveDocument.recompute()
def pageDirtyChanged(self, page):
"""pageDirtyChanged(page) ... internal callback"""
self.buttonBox.button(QtGui.QDialogButtonBox.Apply).setEnabled(self.isDirty())
def clicked(self, button):
"""clicked(button) ... callback invoked when the user presses any of the task panel buttons."""
if button == QtGui.QDialogButtonBox.Apply:
self.panelGetFields()
self.setClean()
FreeCAD.ActiveDocument.recompute()
def modifyStandardButtons(self, buttonBox):
"""modifyStandarButtons(buttonBox) ... callback in case the task panel buttons need to be modified."""
self.buttonBox = buttonBox
for page in self.featurePages:
page.modifyStandardButtons(buttonBox)
self.pageDirtyChanged(None)
def panelGetFields(self):
"""panelGetFields() ... invoked to trigger a complete transfer of UI data to the model."""
Path.Log.track()
for page in self.featurePages:
page.pageGetFields()
def panelSetFields(self):
"""panelSetFields() ... invoked to trigger a complete transfer of the model's properties to the UI."""
Path.Log.track()
self.obj.Proxy.sanitizeBase(self.obj)
for page in self.featurePages:
page.pageSetFields()
def panelCleanup(self):
"""panelCleanup() ... invoked before the receiver is destroyed."""
Path.Log.track()
for page in self.featurePages:
page.pageCleanup()
def open(self):
"""open() ... callback invoked when the task panel is opened."""
self.selectionFactory()
FreeCADGui.Selection.addObserver(self)
def getStandardButtons(self):
"""getStandardButtons() ... returns the Buttons for the task panel."""
return int(
QtGui.QDialogButtonBox.Ok
| QtGui.QDialogButtonBox.Apply
| QtGui.QDialogButtonBox.Cancel
)
def setupUi(self):
"""setupUi() ... internal function to initialise all pages."""
Path.Log.track(self.deleteOnReject)
if (
self.deleteOnReject
and PathOp.FeatureBaseGeometry & self.obj.Proxy.opFeatures(self.obj)
):
sel = FreeCADGui.Selection.getSelectionEx()
for page in self.featurePages:
if getattr(page, "InitBase", True) and hasattr(page, "addBase"):
page.clearBase()
page.addBaseGeometry(sel)
# Update properties based upon expressions in case expression value has changed
for prp, expr in self.obj.ExpressionEngine:
val = FreeCAD.Units.Quantity(self.obj.evalExpression(expr))
value = val.Value if hasattr(val, "Value") else val
prop = getattr(self.obj, prp)
if hasattr(prop, "Value"):
prop.Value = value
else:
prop = value
self.panelSetFields()
for page in self.featurePages:
page.pageRegisterSignalHandlers()
def updateData(self, obj, prop):
"""updateDate(obj, prop) ... callback invoked whenever a model's property is assigned a value."""
# Path.Log.track(obj.Label, prop) # creates a lot of noise
for page in self.featurePages:
page.pageUpdateData(obj, prop)
def needsFullSpace(self):
return True
def updateSelection(self):
sel = FreeCADGui.Selection.getSelectionEx()
for page in self.featurePages:
page.updateSelection(self.obj, sel)
# SelectionObserver interface
def addSelection(self, doc, obj, sub, pnt):
self.updateSelection()
def removeSelection(self, doc, obj, sub):
self.updateSelection()
def setSelection(self, doc):
self.updateSelection()
def clearSelection(self, doc):
self.updateSelection()
class CommandSetStartPoint:
"""Command to set the start point for an operation."""
def GetResources(self):
return {
"Pixmap": "Path_StartPoint",
"MenuText": QT_TRANSLATE_NOOP("PathOp", "Pick Start Point"),
"ToolTip": QT_TRANSLATE_NOOP("PathOp", "Pick Start Point"),
}
def IsActive(self):
if FreeCAD.ActiveDocument is None:
return False
sel = FreeCADGui.Selection.getSelection()
if not sel:
return False
obj = sel[0]
return obj and hasattr(obj, "StartPoint")
def setpoint(self, point, o):
obj = FreeCADGui.Selection.getSelection()[0]
obj.StartPoint.x = point.x
obj.StartPoint.y = point.y
obj.StartPoint.z = obj.ClearanceHeight.Value
def Activated(self):
if not hasattr(FreeCADGui, "Snapper"):
import DraftTools
FreeCADGui.Snapper.getPoint(callback=self.setpoint)
def Create(res):
"""Create(res) ... generic implementation of a create function.
res is an instance of CommandResources. It is not expected that the user invokes
this function directly, but calls the Activated() function of the Command object
that is created in each operations Gui implementation."""
FreeCAD.ActiveDocument.openTransaction("Create %s" % res.name)
try:
obj = res.objFactory(res.name, obj=None, parentJob=res.job)
if obj.Proxy:
obj.ViewObject.Proxy = ViewProvider(obj.ViewObject, res)
obj.ViewObject.Visibility = True
FreeCAD.ActiveDocument.commitTransaction()
obj.ViewObject.Document.setEdit(obj.ViewObject, 0)
return obj
except PathUtils.PathNoTCExistsException:
msg = translate(
"PathOp", "No suitable tool controller found.\nAborting op creation"
)
diag = QtGui.QMessageBox(QtGui.QMessageBox.Warning, "Error", msg)
diag.setWindowModality(QtCore.Qt.ApplicationModal)
diag.exec_()
except PathOp.PathNoTCException:
Path.Log.warning(
translate("PathOp", "No tool controller, aborting op creation")
)
FreeCAD.ActiveDocument.abortTransaction()
FreeCAD.ActiveDocument.recompute()
return None
class CommandPathOp:
"""Generic, data driven implementation of a Path operation creation command.
Instances of this class are stored in all Path operation Gui modules and can
be used to create said operations with view providers and all."""
def __init__(self, resources):
self.res = resources
def GetResources(self):
ress = {
"Pixmap": self.res.pixmap,
"MenuText": self.res.menuText,
"ToolTip": self.res.toolTip,
}
if self.res.accelKey:
ress["Accel"] = self.res.accelKey
return ress
def IsActive(self):
if FreeCAD.ActiveDocument is not None:
for o in FreeCAD.ActiveDocument.Objects:
if o.Name[:3] == "Job":
return True
return False
def Activated(self):
return Create(self.res)
class CommandResources:
"""POD class to hold command specific resources."""
def __init__(
self, name, objFactory, opPageClass, pixmap, menuText, accelKey, toolTip
):
self.name = name
self.objFactory = objFactory
self.opPageClass = opPageClass
self.pixmap = pixmap
self.menuText = menuText
self.accelKey = accelKey
self.toolTip = toolTip
self.job = None
def SetupOperation(
name, objFactory, opPageClass, pixmap, menuText, toolTip, setupProperties=None
):
"""SetupOperation(name, objFactory, opPageClass, pixmap, menuText, toolTip, setupProperties=None)
Creates an instance of CommandPathOp with the given parameters and registers the command with FreeCAD.
When activated it creates a model with proxy (by invoking objFactory), assigns a view provider to it
(see ViewProvider in this module) and starts the editor specifically for this operation (driven by opPageClass).
This is an internal function that is automatically called by the initialisation code for each operation.
It is not expected to be called manually.
"""
res = CommandResources(
name, objFactory, opPageClass, pixmap, menuText, None, toolTip
)
command = CommandPathOp(res)
FreeCADGui.addCommand("Path_%s" % name.replace(" ", "_"), command)
if setupProperties is not None:
PathSetupSheet.RegisterOperation(name, objFactory, setupProperties)
return command
FreeCADGui.addCommand("Path_SetStartPoint", CommandSetStartPoint())
FreeCAD.Console.PrintLog("Loading PathOpGui... done\n")
|
song | songconstants | #!/usr/bin/env python
# -*- coding: utf-8 -*-
#####################################################################
# Frets on Fire #
# Copyright (C) 2006 Sami Kyöstilä #
# 2008 myfingershurt #
# 2009 Pascal Giard #
# #
# This program is free software; you can redistribute it and/or #
# modify it under the terms of the GNU General Public License #
# as published by the Free Software Foundation; either version 2 #
# of the License, or (at your option) any later version. #
# #
# This program is distributed in the hope that it will be useful, #
# but WITHOUT ANY WARRANTY; without even the implied warranty of #
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the #
# GNU General Public License for more details. #
# #
# You should have received a copy of the GNU General Public License #
# along with this program; if not, write to the Free Software #
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, #
# MA 02110-1301, USA. #
#####################################################################
# Song difficulties
EXP_DIF = 0
HAR_DIF = 1
MED_DIF = 2
EAS_DIF = 3
# MIDI note value constants
EXP_GEM_0 = 0x60
EXP_GEM_1 = 0x61
EXP_GEM_2 = 0x62
EXP_GEM_3 = 0x63
EXP_GEM_4 = 0x64
HAR_GEM_0 = 0x54
HAR_GEM_1 = 0x55
HAR_GEM_2 = 0x56
HAR_GEM_3 = 0x57
HAR_GEM_4 = 0x58
MED_GEM_0 = 0x48
MED_GEM_1 = 0x49
MED_GEM_2 = 0x4A
MED_GEM_3 = 0x4B
MED_GEM_4 = 0x4C
EAS_GEM_0 = 0x3C
EAS_GEM_1 = 0x3D
EAS_GEM_2 = 0x3E
EAS_GEM_3 = 0x3F
EAS_GEM_4 = 0x40
#
# One MIDI track is divided into four 'zones' representing the four difficulty
# levels. This dict maps the MIDI notes to difficulty/button pairs. The
# specific notes listed above must be used to indicate which button should be
# pressed.
#
NOTE_MAP = { # difficulty, note
EXP_GEM_0: (EXP_DIF, 0), # ======== #0x60 = 96 = C 8
EXP_GEM_1: (EXP_DIF, 1), # 0x61 = 97 = Db8
EXP_GEM_2: (EXP_DIF, 2), # 0x62 = 98 = D 8
EXP_GEM_3: (EXP_DIF, 3), # 0x63 = 99 = Eb8
EXP_GEM_4: (EXP_DIF, 4), # 0x64 = 100= E 8
HAR_GEM_0: (HAR_DIF, 0), # ======== #0x54 = 84 = C 7
HAR_GEM_1: (HAR_DIF, 1), # 0x55 = 85 = Db7
HAR_GEM_2: (HAR_DIF, 2), # 0x56 = 86 = D 7
HAR_GEM_3: (HAR_DIF, 3), # 0x57 = 87 = Eb7
HAR_GEM_4: (HAR_DIF, 4), # 0x58 = 88 = E 7
MED_GEM_0: (MED_DIF, 0), # ======== #0x48 = 72 = C 6
MED_GEM_1: (MED_DIF, 1), # 0x49 = 73 = Db6
MED_GEM_2: (MED_DIF, 2), # 0x4a = 74 = D 6
MED_GEM_3: (MED_DIF, 3), # 0x4b = 75 = Eb6
MED_GEM_4: (MED_DIF, 4), # 0x4c = 76 = E 6
EAS_GEM_0: (EAS_DIF, 0), # ======== #0x3c = 60 = C 5
EAS_GEM_1: (EAS_DIF, 1), # 0x3d = 61 = Db5
EAS_GEM_2: (EAS_DIF, 2), # 0x3e = 62 = D 5
EAS_GEM_3: (EAS_DIF, 3), # 0x3f = 63 = Eb5
EAS_GEM_4: (EAS_DIF, 4), # 0x40 = 64 = E 5
}
# Real / Pro Guitar midi markers
EXP_REAL_GTR_E2 = 0x60
EXP_REAL_GTR_A2 = 0x61
EXP_REAL_GTR_D3 = 0x62
EXP_REAL_GTR_G3 = 0x63
EXP_REAL_GTR_B3 = 0x64
EXP_REAL_GTR_E4 = 0x65
HAR_REAL_GTR_E2 = 0x48
HAR_REAL_GTR_A2 = 0x49
HAR_REAL_GTR_D3 = 0x4A
HAR_REAL_GTR_G3 = 0x4B
HAR_REAL_GTR_B3 = 0x4C
HAR_REAL_GTR_E4 = 0x4D
MED_REAL_GTR_E2 = 0x30
MED_REAL_GTR_A2 = 0x31
MED_REAL_GTR_D3 = 0x32
MED_REAL_GTR_G3 = 0x33
MED_REAL_GTR_B3 = 0x34
MED_REAL_GTR_E4 = 0x35
EAS_REAL_GTR_E2 = 0x18
EAS_REAL_GTR_A2 = 0x19
EAS_REAL_GTR_D3 = 0x1A
EAS_REAL_GTR_G3 = 0x1B
EAS_REAL_GTR_B3 = 0x1C
EAS_REAL_GTR_E4 = 0x1D
#
# As above for NOTE_MAP, but for the "real" guitar (whatever that is.)
#
REAL_GTR_NOTE_MAP = { # difficulty, note
EXP_REAL_GTR_E2: (EXP_DIF, 0),
EXP_REAL_GTR_A2: (EXP_DIF, 1),
EXP_REAL_GTR_D3: (EXP_DIF, 2),
EXP_REAL_GTR_G3: (EXP_DIF, 3),
EXP_REAL_GTR_B3: (EXP_DIF, 4),
EXP_REAL_GTR_E4: (EXP_DIF, 5),
HAR_REAL_GTR_E2: (HAR_DIF, 0),
HAR_REAL_GTR_A2: (HAR_DIF, 1),
HAR_REAL_GTR_D3: (HAR_DIF, 2),
HAR_REAL_GTR_G3: (HAR_DIF, 3),
HAR_REAL_GTR_B3: (HAR_DIF, 4),
HAR_REAL_GTR_E4: (HAR_DIF, 5),
MED_REAL_GTR_E2: (MED_DIF, 0),
MED_REAL_GTR_A2: (MED_DIF, 1),
MED_REAL_GTR_D3: (MED_DIF, 2),
MED_REAL_GTR_G3: (MED_DIF, 3),
MED_REAL_GTR_B3: (MED_DIF, 4),
MED_REAL_GTR_E4: (MED_DIF, 5),
EAS_REAL_GTR_E2: (EAS_DIF, 0),
EAS_REAL_GTR_A2: (EAS_DIF, 1),
EAS_REAL_GTR_D3: (EAS_DIF, 2),
EAS_REAL_GTR_G3: (EAS_DIF, 3),
EAS_REAL_GTR_B3: (EAS_DIF, 4),
EAS_REAL_GTR_E4: (EAS_DIF, 5),
}
# -- special note numbers -- #
# Star Power
SP_MARKING_NOTE = 0x67 # note 103 = G 8
# Overdrive
OD_MARKING_NOTE = 0x74 # note 116 = G#9
# Hammer-On / Pull-Off
HOPO_MARKING_NOTES = []
PRO_HOPO_MARKING_NOTES = [0x1E, 0x36, 0x4E, 0x66]
# Free Style
FREESTYLE_MARKING_NOTE = (
0x7C # notes 120 - 124 = drum fills & BREs - always all 5 notes present
)
DEFAULT_BPM = 120.0
DEFAULT_LIBRARY = "songs"
# special / global text-event tracks for the separated text-event list variable
TK_SCRIPT = 0 # script.txt events
TK_SECTIONS = 1 # Section events
TK_GUITAR_SOLOS = 2 # Guitar Solo start/stop events
TK_LYRICS = 3 # RB MIDI Lyric events
TK_UNUSED_TEXT = 4 # Unused / other text events
GUITAR_TRACK = 0
RHYTHM_TRACK = 1
DRUM_TRACK = 2
VOCAL_TRACK = 3
MIDI_TYPE_GH = 0 # GH type MIDIs have starpower phrases marked with a long G8 note on that instrument's track
MIDI_TYPE_RB = 1 # RB type MIDIs have overdrive phrases marked with a long G#9 note on that instrument's track
MIDI_TYPE_WT = 2 # WT type MIDIs have six notes and HOPOs marked on F# of the track
EARLY_HIT_WINDOW_NONE = 1 # GH1/RB1/RB2 = NONE
EARLY_HIT_WINDOW_HALF = 2 # GH2/GH3/GHA/GH80's = HALF
EARLY_HIT_WINDOW_FULL = 3 # FoF = FULL
GUITAR_PART = 0
RHYTHM_PART = 1
BASS_PART = 2
LEAD_PART = 3
DRUM_PART = 4
VOCAL_PART = 5
KEYS_PART = 6
PRO_GUITAR_PART = 7
PRO_BASS_PART = 8
PRO_DRUM_PART = 9
PRO_KEYS_PART = 10
|
protos | string_int_label_map_pb2 | # Generated by the protocol buffer compiler. DO NOT EDIT!
# source: object_detection/protos/string_int_label_map.proto
import sys
_b = sys.version_info[0] < 3 and (lambda x: x) or (lambda x: x.encode("latin1"))
from google.protobuf import descriptor as _descriptor
from google.protobuf import descriptor_pb2
from google.protobuf import message as _message
from google.protobuf import reflection as _reflection
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
DESCRIPTOR = _descriptor.FileDescriptor(
name="object_detection/protos/string_int_label_map.proto",
package="object_detection.protos",
syntax="proto2",
serialized_pb=_b(
'\n2object_detection/protos/string_int_label_map.proto\x12\x17object_detection.protos"G\n\x15StringIntLabelMapItem\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\n\n\x02id\x18\x02 \x01(\x05\x12\x14\n\x0c\x64isplay_name\x18\x03 \x01(\t"Q\n\x11StringIntLabelMap\x12<\n\x04item\x18\x01 \x03(\x0b\x32..object_detection.protos.StringIntLabelMapItem'
),
)
_STRINGINTLABELMAPITEM = _descriptor.Descriptor(
name="StringIntLabelMapItem",
full_name="object_detection.protos.StringIntLabelMapItem",
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name="name",
full_name="object_detection.protos.StringIntLabelMapItem.name",
index=0,
number=1,
type=9,
cpp_type=9,
label=1,
has_default_value=False,
default_value=_b("").decode("utf-8"),
message_type=None,
enum_type=None,
containing_type=None,
is_extension=False,
extension_scope=None,
options=None,
file=DESCRIPTOR,
),
_descriptor.FieldDescriptor(
name="id",
full_name="object_detection.protos.StringIntLabelMapItem.id",
index=1,
number=2,
type=5,
cpp_type=1,
label=1,
has_default_value=False,
default_value=0,
message_type=None,
enum_type=None,
containing_type=None,
is_extension=False,
extension_scope=None,
options=None,
file=DESCRIPTOR,
),
_descriptor.FieldDescriptor(
name="display_name",
full_name="object_detection.protos.StringIntLabelMapItem.display_name",
index=2,
number=3,
type=9,
cpp_type=9,
label=1,
has_default_value=False,
default_value=_b("").decode("utf-8"),
message_type=None,
enum_type=None,
containing_type=None,
is_extension=False,
extension_scope=None,
options=None,
file=DESCRIPTOR,
),
],
extensions=[],
nested_types=[],
enum_types=[],
options=None,
is_extendable=False,
syntax="proto2",
extension_ranges=[],
oneofs=[],
serialized_start=79,
serialized_end=150,
)
_STRINGINTLABELMAP = _descriptor.Descriptor(
name="StringIntLabelMap",
full_name="object_detection.protos.StringIntLabelMap",
filename=None,
file=DESCRIPTOR,
containing_type=None,
fields=[
_descriptor.FieldDescriptor(
name="item",
full_name="object_detection.protos.StringIntLabelMap.item",
index=0,
number=1,
type=11,
cpp_type=10,
label=3,
has_default_value=False,
default_value=[],
message_type=None,
enum_type=None,
containing_type=None,
is_extension=False,
extension_scope=None,
options=None,
file=DESCRIPTOR,
),
],
extensions=[],
nested_types=[],
enum_types=[],
options=None,
is_extendable=False,
syntax="proto2",
extension_ranges=[],
oneofs=[],
serialized_start=152,
serialized_end=233,
)
_STRINGINTLABELMAP.fields_by_name["item"].message_type = _STRINGINTLABELMAPITEM
DESCRIPTOR.message_types_by_name["StringIntLabelMapItem"] = _STRINGINTLABELMAPITEM
DESCRIPTOR.message_types_by_name["StringIntLabelMap"] = _STRINGINTLABELMAP
_sym_db.RegisterFileDescriptor(DESCRIPTOR)
StringIntLabelMapItem = _reflection.GeneratedProtocolMessageType(
"StringIntLabelMapItem",
(_message.Message,),
dict(
DESCRIPTOR=_STRINGINTLABELMAPITEM,
__module__="object_detection.protos.string_int_label_map_pb2",
# @@protoc_insertion_point(class_scope:object_detection.protos.StringIntLabelMapItem)
),
)
_sym_db.RegisterMessage(StringIntLabelMapItem)
StringIntLabelMap = _reflection.GeneratedProtocolMessageType(
"StringIntLabelMap",
(_message.Message,),
dict(
DESCRIPTOR=_STRINGINTLABELMAP,
__module__="object_detection.protos.string_int_label_map_pb2",
# @@protoc_insertion_point(class_scope:object_detection.protos.StringIntLabelMap)
),
)
_sym_db.RegisterMessage(StringIntLabelMap)
# @@protoc_insertion_point(module_scope)
|
config-integrations | formatted_webhook | # Main
enabled = True
title = "Formatted webhook"
slug = "formatted_webhook"
short_description = None
description = None
is_displayed_on_web = True
is_featured = False
is_able_to_autoresolve = True
is_demo_alert_enabled = True
description = None
# Default templates
slack_title = """\
*<{{ grafana_oncall_link }}|#{{ grafana_oncall_incident_id }} {{ payload.get("title", "Title undefined (Check Slack Title Template)") }}>* via {{ integration_name }}
{% if source_link %}
(*<{{ source_link }}|source>*)
{%- endif %}"""
slack_message = "{{ payload.message }}"
slack_image_url = "{{ payload.image_url }}"
web_title = '{{ payload.get("title", "Title undefined (Check Web Title Template)") }}'
web_message = slack_message
web_image_url = slack_image_url
sms_title = web_title
phone_call_title = sms_title
telegram_title = sms_title
telegram_message = slack_message
telegram_image_url = slack_image_url
source_link = "{{ payload.link_to_upstream_details }}"
grouping_id = '{{ payload.get("alert_uid", "") }}'
resolve_condition = '{{ payload.get("state", "").upper() == "OK" }}'
acknowledge_condition = None
example_payload = {
"alert_uid": "08d6891a-835c-e661-39fa-96b6a9e26552",
"title": "TestAlert: The whole system is down",
"image_url": "https://upload.wikimedia.org/wikipedia/commons/e/ee/Grumpy_Cat_by_Gage_Skidmore.jpg",
"state": "alerting",
"link_to_upstream_details": "https://en.wikipedia.org/wiki/Downtime",
"message": "This alert was sent by user for demonstration purposes\nSmth happened. Oh no!",
}
|
PyObjCTest | test_cfplugin | """
No tests for CFPlugin: these API's are unsupported for the moment.
"""
import CoreFoundation
from PyObjCTools.TestSupport import *
symbols = [
# From CFPluginCOM.h:
"IUnknownVTbl",
"IS_ERROR",
"HRESULT_CODE",
"HRESULT_FACILITY",
"SEVERITY_SUCCESS",
"SEVERITY_ERROR",
"MAKE_HRESULT",
"S_OK",
"S_FALSE",
"E_UNEXPECTED",
"E_NOTIMPL",
"E_OUTOFMEMORY",
"E_INVALIDARG",
"E_NOINTERFACE",
"E_POINTER",
"E_HANDLE",
"E_ABORT",
"E_FAIL",
"E_ACCESSDENIED",
# From CFPlugin.h:
"kCFPlugInDynamicRegistrationKey",
"kCFPlugInDynamicRegisterFunctionKey",
"kCFPlugInUnloadFunctionKey",
"kCFPlugInFactoriesKey",
"kCFPlugInTypesKey",
"CFPlugInGetTypeID",
"CFPlugInCreate",
"CFPlugInGetBundle",
"CFPlugInSetLoadOnDemand",
"CFPlugInIsLoadOnDemand",
"CFPlugInFindFactoriesForPlugInType",
"CFPlugInFindFactoriesForPlugInTypeInPlugIn",
"CFPlugInInstanceCreate",
"CFPlugInRegisterFactoryFunction",
"CFPlugInRegisterFactoryFunctionByName",
"CFPlugInUnregisterFactory",
"CFPlugInRegisterPlugInType",
"CFPlugInUnregisterPlugInType",
"CFPlugInAddInstanceForFactory",
"CFPlugInRemoveInstanceForFactory",
"CFPlugInInstanceGetInterfaceFunctionTable",
"CFPlugInInstanceGetFactoryName",
"CFPlugInInstanceGetInstanceData",
"CFPlugInInstanceGetTypeID",
"CFPlugInInstanceCreateWithInstanceDataSize",
]
class TestPluginNotSuppported(TestCase):
def testUnsupported(self):
for sym in symbols:
if hasattr(CoreFoundation, sym):
self.fail("Unsupported symbol present: %s" % (sym,))
if __name__ == "__main__":
main()
|
cloudfiles | engine | # coding:utf-8
import pyrax
from cactus.deployment.cloudfiles.auth import CloudFilesCredentialsManager
from cactus.deployment.cloudfiles.file import CloudFilesFile
from cactus.deployment.engine import BaseDeploymentEngine
class CloudFilesDeploymentEngine(BaseDeploymentEngine):
CredentialsManagerClass = CloudFilesCredentialsManager
FileClass = CloudFilesFile
config_bucket_name = "cloudfiles-bucket-name"
config_bucket_website = "cloudfiles-bucket-website"
def _create_connection(self):
username, api_key = self.credentials_manager.get_credentials()
pyrax.set_setting("identity_type", "rackspace")
pyrax.set_credentials(username, api_key)
return pyrax.connect_to_cloudfiles()
def get_bucket(self):
try:
return self.get_connection().get_container(self.bucket_name)
except pyrax.exceptions.NoSuchContainer:
return None
def create_bucket(self):
# TODO: Handle errors
conn = self.get_connection()
container = conn.create_container(self.bucket_name)
container.set_web_index_page(self._index_page)
container.set_web_error_page(self._error_page)
container.make_public()
return container
def get_website_endpoint(self):
return self.bucket.cdn_uri
|
builders | box_coder_builder | # Copyright 2017 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
"""A function to build an object detection box coder from configuration."""
from app.object_detection.box_coders import (
faster_rcnn_box_coder,
keypoint_box_coder,
mean_stddev_box_coder,
square_box_coder,
)
from app.object_detection.protos import box_coder_pb2
def build(box_coder_config):
"""Builds a box coder object based on the box coder config.
Args:
box_coder_config: A box_coder.proto object containing the config for the
desired box coder.
Returns:
BoxCoder based on the config.
Raises:
ValueError: On empty box coder proto.
"""
if not isinstance(box_coder_config, box_coder_pb2.BoxCoder):
raise ValueError("box_coder_config not of type box_coder_pb2.BoxCoder.")
if box_coder_config.WhichOneof("box_coder_oneof") == "faster_rcnn_box_coder":
return faster_rcnn_box_coder.FasterRcnnBoxCoder(
scale_factors=[
box_coder_config.faster_rcnn_box_coder.y_scale,
box_coder_config.faster_rcnn_box_coder.x_scale,
box_coder_config.faster_rcnn_box_coder.height_scale,
box_coder_config.faster_rcnn_box_coder.width_scale,
]
)
if box_coder_config.WhichOneof("box_coder_oneof") == "keypoint_box_coder":
return keypoint_box_coder.KeypointBoxCoder(
box_coder_config.keypoint_box_coder.num_keypoints,
scale_factors=[
box_coder_config.keypoint_box_coder.y_scale,
box_coder_config.keypoint_box_coder.x_scale,
box_coder_config.keypoint_box_coder.height_scale,
box_coder_config.keypoint_box_coder.width_scale,
],
)
if box_coder_config.WhichOneof("box_coder_oneof") == "mean_stddev_box_coder":
return mean_stddev_box_coder.MeanStddevBoxCoder()
if box_coder_config.WhichOneof("box_coder_oneof") == "square_box_coder":
return square_box_coder.SquareBoxCoder(
scale_factors=[
box_coder_config.square_box_coder.y_scale,
box_coder_config.square_box_coder.x_scale,
box_coder_config.square_box_coder.length_scale,
]
)
raise ValueError("Empty box coder.")
|
classes | thumbnail | """
@file
@brief This file has code to generate thumbnail images and HTTP thumbnail server
@author Jonathan Thomas <jonathan@openshot.org>
@section LICENSE
Copyright (c) 2008-2018 OpenShot Studios, LLC
(http://www.openshotstudios.com). This file is part of
OpenShot Video Editor (http://www.openshot.org), an open-source project
dedicated to delivering high quality video editing and animation solutions
to the world.
OpenShot Video Editor is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
OpenShot Video Editor is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with OpenShot Library. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import re
import socket
import time
from http.server import BaseHTTPRequestHandler, HTTPServer
from socketserver import ThreadingMixIn
from threading import Thread
import openshot
from classes import info
from classes.app import get_app
from classes.logger import log
from classes.query import File
from requests import get
# Regex for parsing URLs: (examples)
# http://127.0.0.1:33723/thumbnails/9ATJTBQ71V/1/path/no-cache/
# http://127.0.0.1:33723/thumbnails/9ATJTBQ71V/1/path/
# http://127.0.0.1:33723/thumbnails/9ATJTBQ71V/1/path
# http://127.0.0.1:33723/thumbnails/9ATJTBQ71V/1/
# http://127.0.0.1:33723/thumbnails/9ATJTBQ71V/1
REGEX_THUMBNAIL_URL = re.compile(
r"/thumbnails/(?P<file_id>.+?)/(?P<file_frame>\d+)/*(?P<only_path>path)?/*(?P<no_cache>no-cache)?"
)
def GetThumbPath(file_id, thumbnail_frame, clear_cache=False):
"""Get thumbnail path by invoking HTTP thumbnail request"""
# Clear thumb cache (if requested)
thumb_cache = ""
if clear_cache:
thumb_cache = "no-cache/"
# Connect to thumbnail server and get image
thumb_server_details = get_app().window.http_server_thread.server_address
thumb_address = "http://%s:%s/thumbnails/%s/%s/path/%s" % (
thumb_server_details[0],
thumb_server_details[1],
file_id,
thumbnail_frame,
thumb_cache,
)
r = get(thumb_address)
if r.ok:
# Update thumbnail path to real one
return r.text
else:
return ""
def GenerateThumbnail(
file_path, thumb_path, thumbnail_frame, width, height, mask, overlay
):
"""Create thumbnail image, and check for rotate metadata (if any)"""
# Create a clip object and get the reader
clip = openshot.Clip(file_path)
reader = clip.Reader()
scale = get_app().devicePixelRatio()
if scale > 1.0:
clip.scale_x.AddPoint(1.0, 1.0 * scale)
clip.scale_y.AddPoint(1.0, 1.0 * scale)
# Open reader
reader.Open()
# Get the 'rotate' metadata (if any)
rotate = 0.0
try:
if reader.info.metadata.count("rotate"):
rotate_data = reader.info.metadata.find("rotate").value()[1]
rotate = float(rotate_data)
except ValueError as ex:
log.warning("Could not parse rotation value {}: {}".format(rotate_data, ex))
except Exception:
log.warning(
"Error reading rotation metadata from {}".format(file_path), exc_info=1
)
# Create thumbnail folder (if needed)
parent_path = os.path.dirname(thumb_path)
if not os.path.exists(parent_path):
os.mkdir(parent_path)
# Save thumbnail image and close readers
reader.GetFrame(thumbnail_frame).Thumbnail(
thumb_path,
round(width * scale),
round(height * scale),
mask,
overlay,
"#000",
False,
"png",
85,
rotate,
)
reader.Close()
clip.Close()
class httpThumbnailServer(ThreadingMixIn, HTTPServer):
"""This class allows to handle requests in separated threads.
No further content needed, don't touch this."""
class httpThumbnailException(Exception):
"""Custom exception if server cannot start. This can happen if a port does ot allow a connection
due to another program or due to a firewall."""
class httpThumbnailServerThread(Thread):
"""This class runs a HTTP thumbnail server inside a thread
so we don't block the main thread with handle_request()."""
def find_free_port(self):
"""Find the first available socket port"""
s = socket.socket()
s.bind(("", 0))
socket_port = s.getsockname()[1]
s.close()
return socket_port
def kill(self):
self.running = False
log.info("Shutting down thumbnail server: %s" % str(self.server_address))
self.thumbServer.shutdown()
def run(self):
log.info("Starting thumbnail server listening on %s", self.server_address)
self.running = True
self.thumbServer.serve_forever(0.5)
def __init__(self):
"""Attempt to find an available port, and bind to that port for our thumbnail HTTP server.
If not able to bind to localhost or a specific port, return an exception (and quit OpenShot).
"""
Thread.__init__(self)
self.daemon = True
self.server_address = None
self.running = False
self.thumbServer = None
exceptions = []
initial_port = self.find_free_port()
for attempt in range(3):
try:
# Configure server address and port for our HTTP thumbnail server
self.server_address = ("127.0.0.1", initial_port + attempt)
log.debug(
"Attempting to start thumbnail server listening on port %s",
self.server_address,
)
self.thumbServer = httpThumbnailServer(
self.server_address, httpThumbnailHandler
)
self.thumbServer.daemon_threads = True
exceptions.clear()
break
except Exception as ex:
# Silently track each exception
# Return full list of exceptions (from each attempt, if no attempt is successful)
exceptions.append(f"{self.server_address} {ex}")
if exceptions:
# Return full list of attempts + exceptions if we failed to make a connection
raise httpThumbnailException("\n".join(exceptions))
class httpThumbnailHandler(BaseHTTPRequestHandler):
"""This class handles HTTP requests to the HTTP thumbnail server above."""
def log_message(self, msg_format, *args):
"""Log message from HTTPServer"""
log.info(msg_format % args)
def log_error(self, msg_format, *args):
"""Log error from HTTPServer"""
log.warning(msg_format % args)
def do_GET(self):
"""Process each GET request and return a value (image or file path)"""
mask_path = os.path.join(info.IMAGES_PATH, "mask.png")
# Parse URL
url_output = REGEX_THUMBNAIL_URL.match(self.path)
if url_output and len(url_output.groups()) == 4:
# Path is expected to have 3 matched components (third is optional though)
# /thumbnails/FILE-ID/FRAME-NUMBER/ or
# /thumbnails/FILE-ID/FRAME-NUMBER/path/ or
# /thumbnails/FILE-ID/FRAME-NUMBER/no-cache/ or
# /thumbnails/FILE-ID/FRAME-NUMBER/path/no-cache/
self.send_response_only(200)
else:
self.send_error(404)
return
# Get URL parts
file_id = url_output.group("file_id")
file_frame = int(url_output.group("file_frame"))
only_path = url_output.group("only_path")
no_cache = url_output.group("no_cache")
log.debug("Processing thumbnail request for %s frame %d", file_id, file_frame)
try:
# Look up file data
file = File.get(id=file_id)
# Ensure file location is an absolute path
file_path = file.absolute_path()
except AttributeError:
# Couldn't match file ID
log.debug("No ID match, returning 404")
self.send_error(404)
return
# Send headers
if not only_path:
self.send_header("Content-type", "image/png")
else:
self.send_header("Content-type", "text/html; charset=utf-8")
self.end_headers()
# Locate thumbnail
thumb_path = os.path.join(info.THUMBNAIL_PATH, file_id, "%s.png" % file_frame)
if not os.path.exists(thumb_path) and file_frame == 1:
# Try ID with no frame # (for backwards compatibility)
thumb_path = os.path.join(info.THUMBNAIL_PATH, "%s.png" % file_id)
if not os.path.exists(thumb_path) and file_frame != 1:
# Try with ID and frame # in filename (for backwards compatibility)
thumb_path = os.path.join(
info.THUMBNAIL_PATH, "%s-%s.png" % (file_id, file_frame)
)
if not os.path.exists(thumb_path) or no_cache:
# Generate thumbnail (since we can't find it)
# Determine if video overlay should be applied to thumbnail
overlay_path = ""
if file.data["media_type"] == "video":
overlay_path = os.path.join(info.IMAGES_PATH, "overlay.png")
# Create thumbnail image
GenerateThumbnail(
file_path, thumb_path, file_frame, 98, 64, mask_path, overlay_path
)
# Send message back to client
if os.path.exists(thumb_path):
if only_path:
self.wfile.write(bytes(thumb_path, "utf-8"))
else:
with open(thumb_path, "rb") as f:
self.wfile.write(f.read())
# Pause processing of request (since we don't currently use thread pooling, this allows
# the threads to be processed without choking the CPU as much
# TODO: Make HTTPServer work with a limited thread pool and remove this sleep() hack.
time.sleep(0.01)
|
gui | linemode | # This file is part of MyPaint.
# -*- coding: utf-8 -*-
# Copyright (C) 2012 by Richard Jones
# Copyright (C) 2012-2018 by the MyPaint Development Team.
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or
# (at your option) any later version.
# NOTE: Much of this was written before the new modes system.
# NOTE: The InteractionMode stuff was sort of bolted on after the fact.
# TODO: Understand it more; devolve specifics to specific subclasses.
# TODO: Document stuff properly.
## Imports
from __future__ import division, print_function
import logging
import math
from gettext import gettext as _
import gui.cursor
import gui.mode
from lib.gibindings import Gdk, GLib, Gtk
from lib.pycompat import xrange
from .curve import CurveWidget
logger = logging.getLogger(__name__)
## Module constants
# internal-name, display-name, constant, minimum, default, maximum, tooltip
_LINE_MODE_SETTINGS_LIST = [
[
"entry_pressure",
_("Entrance Pressure"),
False,
0.0001,
0.3,
1.0,
_("Stroke entrance pressure for line tools"),
],
[
"midpoint_pressure",
_("Midpoint Pressure"),
False,
0.0001,
1.0,
1.0,
_("Mid-Stroke pressure for line tools"),
],
[
"exit_pressure",
_("Exit Pressure"),
False,
0.0001,
0.3,
1.0,
_("Stroke exit pressure for line tools"),
],
["line_head", _("Head"), False, 0.0001, 0.25, 1.0, _("Stroke lead-in end")],
["line_tail", _("Tail"), False, 0.0001, 0.75, 1.0, _("Stroke trail-off beginning")],
]
## Line pressure settings
class LineModeSettings(object):
"""Manage GtkAdjustments for tweaking LineMode settings.
An instance resides in the main application singleton. Changes to the
adjustments are reflected into the app preferences.
"""
def __init__(self, app):
"""Initializer; initial settings are loaded from the app prefs"""
object.__init__(self)
self.app = app
self.adjustments = {} #: Dictionary of GtkAdjustments
self.observers = [] #: List of callbacks
self._idle_srcid = None
self._changed_settings = set()
for line_list in _LINE_MODE_SETTINGS_LIST:
cname, name, const, min_, default, max_, tooltip = line_list
prefs_key = "linemode.%s" % cname
value = float(self.app.preferences.get(prefs_key, default))
adj = Gtk.Adjustment(
value=value,
lower=min_,
upper=max_,
step_increment=0.01,
page_increment=0.1,
)
adj.connect("value-changed", self._value_changed_cb, prefs_key)
self.adjustments[cname] = adj
def _value_changed_cb(self, adj, prefs_key):
# Direct GtkAdjustment callback for a single adjustment being changed.
value = float(adj.get_value())
self.app.preferences[prefs_key] = value
self._changed_settings.add(prefs_key)
if self._idle_srcid is None:
self._idle_srcid = GLib.idle_add(self._values_changed_idle_cb)
def _values_changed_idle_cb(self):
# Aggregate, idle-state callback for multiple adjustments being changed
# in a single event. Queues redraws, and runs observers. The curve sets
# multiple settings at once, and we might as well not queue too many
# redraws.
if self._idle_srcid is not None:
current_mode = self.app.doc.modes.top
if isinstance(current_mode, LineModeBase):
# Redraw last_line when settings are adjusted
# in the adjustment Curve
GLib.idle_add(current_mode.redraw_line_cb)
for func in self.observers:
func(self._changed_settings)
self._changed_settings = set()
self._idle_srcid = None
return False
class LineModeCurveWidget(CurveWidget):
"""Graph of pressure by distance, tied to the central LineModeSettings"""
_SETTINGS_COORDINATE = [
("entry_pressure", (0, 1)),
("midpoint_pressure", (1, 1)),
("exit_pressure", (3, 1)),
("line_head", (1, 0)),
("line_tail", (2, 0)),
]
def __init__(self):
from gui.application import get_app
self.app = get_app()
CurveWidget.__init__(
self, npoints=4, ylockgroups=((1, 2),), changed_cb=self._changed_cb
)
self.app.line_mode_settings.observers.append(self._adjs_changed_cb)
self._update()
def _adjs_changed_cb(self, changed):
logger.debug("Updating curve (changed: %r)", changed)
self._update()
def _update(self, from_defaults=False):
if from_defaults:
self.points = [(0.0, 0.2), (0.33, 0.5), (0.66, 0.5), (1.0, 0.33)]
for setting, coord_pair in self._SETTINGS_COORDINATE:
if not from_defaults:
adj = self.app.line_mode_settings.adjustments[setting]
value = adj.get_value()
else:
# TODO: move this case into the base settings object
defaults = [a[4] for a in _LINE_MODE_SETTINGS_LIST if a[0] == setting]
assert len(defaults) == 1
value = defaults[0]
index, subindex = coord_pair
if not setting.startswith("line"):
value = 1.0 - value
if subindex == 0:
coord = (value, self.points[index][1])
else:
coord = (self.points[index][0], value)
self.set_point(index, coord)
if from_defaults:
self._changed_cb(self)
self.queue_draw()
def _changed_cb(self, curve):
"""Updates the linemode pressure settings when the curve is altered"""
for setting, coord_pair in self._SETTINGS_COORDINATE:
index, subindex = coord_pair
value = self.points[index][subindex]
if not setting.startswith("line"):
value = 1.0 - value
value = max(0.0001, value)
adj = self.app.line_mode_settings.adjustments[setting]
adj.set_value(value)
## Options UI
class LineModeOptionsWidget(gui.mode.PaintingModeOptionsWidgetBase):
"""Options widget for geometric line modes"""
def init_specialized_widgets(self, row=0):
curve = LineModeCurveWidget()
curve.set_size_request(175, 125)
self._curve = curve
exp = Gtk.Expander()
exp.set_label(_("Pressure variation…"))
exp.set_use_markup(False)
exp.add(curve)
self.attach(exp, 0, row, 2, 1)
row += 1
return row
def reset_button_clicked_cb(self, button):
super(LineModeOptionsWidget, self).reset_button_clicked_cb(button)
self._curve._update(from_defaults=True)
## Interaction modes for making lines
class LineModeBase(
gui.mode.ScrollableModeMixin, gui.mode.BrushworkModeMixin, gui.mode.DragMode
):
"""Draws geometric lines (base class)"""
## Class constants
_OPTIONS_WIDGET = None
## Class configuration.
permitted_switch_actions = {
"PanViewMode",
"ZoomViewMode",
"RotateViewMode",
"BrushResizeMode",
}.union(gui.mode.BUTTON_BINDING_ACTIONS)
pointer_behavior = gui.mode.Behavior.PAINT_CONSTRAINED
scroll_behavior = gui.mode.Behavior.CHANGE_VIEW
@property
def active_cursor(self):
cursor_name = gui.cursor.Name.PENCIL
if not self._line_possible:
cursor_name = gui.cursor.Name.FORBIDDEN_EVERYWHERE
return self.doc.app.cursors.get_action_cursor(self.ACTION_NAME, cursor_name)
@classmethod
def get_name(cls):
return _("Lines and Curves")
def get_usage(self):
# TRANSLATORS: users should never see this message
return _("Generic line/curve mode")
@property
def inactive_cursor(self):
cursor_name = gui.cursor.Name.CROSSHAIR_OPEN_PRECISE
if not self._line_possible:
cursor_name = gui.cursor.Name.FORBIDDEN_EVERYWHERE
return self.doc.app.cursors.get_action_cursor(self.ACTION_NAME, cursor_name)
unmodified_persist = True
# FIXME: all of the logic resides in the base class, for historical
# reasons, and is decided by line_mode. The differences should be
# factored out to the user-facing mode subclasses at some point.
line_mode = None
## Initialization
def __init__(self, **kwds):
"""Initialize"""
super(LineModeBase, self).__init__(**kwds)
self.app = None
self.last_line_data = None
self.idle_srcid = None
self._line_possible = False
## InteractionMode/DragMode implementation
def enter(self, doc, **kwds):
"""Enter the mode.
If modifiers are held when the mode is entered, the mode is a oneshot
mode and is popped from the mode stack automatically at the end of the
drag. Without modifiers, line modes may be continued, and some
subclasses offer additional options for adjusting control points.
"""
super(LineModeBase, self).enter(doc, **kwds)
self.app = self.doc.app
rootstack = self.doc.model.layer_stack
rootstack.current_path_updated += self._update_cursors
rootstack.layer_properties_changed += self._update_cursors
self._update_cursors()
def leave(self, **kwds):
rootstack = self.doc.model.layer_stack
rootstack.current_path_updated -= self._update_cursors
rootstack.layer_properties_changed -= self._update_cursors
return super(LineModeBase, self).leave(**kwds)
def _update_cursors(self, *_ignored):
if self.in_drag:
return # defer update to the end of the drag
layer = self.doc.model.layer_stack.current
self._line_possible = layer.get_paintable()
self.doc.tdw.set_override_cursor(self.inactive_cursor)
def drag_start_cb(self, tdw, event):
super(LineModeBase, self).drag_start_cb(tdw, event)
if self._line_possible:
self.start_command(self.initial_modifiers)
def drag_update_cb(self, tdw, event, ev_x, ev_y, dx, dy):
if self._line_possible:
self.update_position(ev_x, ev_y)
if self.idle_srcid is None:
self.idle_srcid = GLib.idle_add(self._drag_idle_cb)
return super(LineModeBase, self).drag_update_cb(tdw, event, ev_x, ev_y, dx, dy)
def drag_stop_cb(self, tdw):
if self._line_possible:
self.idle_srcid = None
self.stop_command()
self._update_cursors()
return super(LineModeBase, self).drag_stop_cb(tdw)
def _drag_idle_cb(self):
# Updates the on-screen line during drags.
if self.idle_srcid is not None:
self.idle_srcid = None
self.process_line()
def checkpoint(self, flush=True, **kwargs):
# Only push outstanding changes to the document's undo stack on
# a request for a flushing sync. Without this, users can't curve
# the previously drawn line. See also inkmode.py and
# https://github.com/mypaint/mypaint/issues/262
if flush:
super(LineModeBase, self).checkpoint(flush=True, **kwargs)
## Options panel
def get_options_widget(self):
"""Get the (base class singleton) options widget"""
cls = LineModeBase
if cls._OPTIONS_WIDGET is None:
widget = LineModeOptionsWidget()
cls._OPTIONS_WIDGET = widget
return cls._OPTIONS_WIDGET
### Draw dynamic Line, Curve, or Ellipse
def start_command(self, modifier):
# :param modifier: the keyboard modifiers which ere in place
# when the mode was created
active_tdw = self.app.doc.tdw.__class__.get_active_tdw()
assert active_tdw is not None
if active_tdw is self.app.scratchpad_doc.tdw:
self.model = self.app.scratchpad_doc.model
self.tdw = self.app.scratchpad_doc.tdw
else:
# unconditionally
self.model = self.app.doc.model
self.tdw = self.app.doc.tdw
self.done = False
self.brushwork_begin(self.model, abrupt=False, description=self.get_name())
layer = self.model.layer_stack.current
x, y, kbmods = self.local_mouse_state()
# ignore the modifier used to start this action (don't make it
# change the action)
self.invert_kbmods = modifier
kbmods ^= self.invert_kbmods # invert using bitwise xor
shift = kbmods & Gdk.ModifierType.SHIFT_MASK
# line_mode is the type of line to be drawn eg. "EllipseMode"
self.mode = self.line_mode
assert self.mode is not None
# Ignore slow_tracking.
# There are some other sttings that interfere
# with the workings of the Line Tools,
# but slowtracking is the main one.
self.adj = self.app.brush_adjustment["slow_tracking"]
self.slow_tracking = self.adj.get_value()
self.adj.set_value(0)
# Throughout this module these conventions are used:
# sx, sy = starting point
# ex, ey = end point
# kx, ky = curve point from last line
# lx, ly = last point from InteractionMode update
self.sx, self.sy = x, y
self.lx, self.ly = x, y
if self.mode == "EllipseMode":
# Rotation angle of ellipse.
self.angle = 90
# Vector to measure any rotation from.
# Assigned when ratation begins.
self.ellipse_vec = None
return
# If not Ellipse, command must be Straight Line or Sequence
# First check if the user intends to Curve or move an existing Line
if shift:
last_line = self.last_line_data
last_stroke = layer.get_last_stroke_info()
if last_line is not None:
if last_line[1] == last_stroke:
self.mode = last_line[0]
self.sx, self.sy = last_line[2], last_line[3]
self.ex, self.ey = last_line[4], last_line[5]
if self.mode == "CurveLine2":
length_a = distance(x, y, self.sx, self.sy)
length_b = distance(x, y, self.ex, self.ey)
self.flip = length_a > length_b
if self.flip:
self.kx, self.ky = last_line[6], last_line[7]
self.k2x, self.k2y = last_line[8], last_line[9]
else:
self.k2x, self.k2y = last_line[6], last_line[7]
self.kx, self.ky = last_line[8], last_line[9]
self.model.undo()
self.process_line()
return
if self.mode == "SequenceMode":
if not self.tdw.last_painting_pos:
return
else:
self.sx, self.sy = self.tdw.last_painting_pos
def update_position(self, x, y):
self.lx, self.ly = self.tdw.display_to_model(x, y)
def stop_command(self):
# End mode
self.done = True
x, y = self.process_line()
self.brushwork_commit(self.model, abrupt=False)
cmd = self.mode
self.record_last_stroke(cmd, x, y)
def record_last_stroke(self, cmd, x, y):
"""Store last stroke data
Stroke data is used for redraws and modifications of the line.
:param str cmd: name of the last command
:param int x: last cursor x-coordinate
:param int y: last cursor y-coordinate
"""
last_line = None
self.tdw.last_painting_pos = x, y
# FIXME: should probably not set that from here
layer = self.model.layer_stack.current
last_stroke = layer.get_last_stroke_info()
sx, sy = self.sx, self.sy
if cmd == "CurveLine1":
last_line = [
"CurveLine2",
last_stroke,
sx,
sy,
self.ex,
self.ey,
x,
y,
x,
y,
]
self.tdw.last_painting_pos = self.ex, self.ey
if cmd == "CurveLine2":
if self.flip:
last_line = [
cmd,
last_stroke,
sx,
sy,
self.ex,
self.ey,
self.kx,
self.ky,
self.k2x,
self.k2y,
]
else:
last_line = [
cmd,
last_stroke,
sx,
sy,
self.ex,
self.ey,
self.k2x,
self.k2y,
self.kx,
self.ky,
]
self.tdw.last_painting_pos = self.ex, self.ey
if cmd == "StraightMode" or cmd == "SequenceMode":
last_line = ["CurveLine1", last_stroke, sx, sy, x, y]
if cmd == "EllipseMode":
last_line = [cmd, last_stroke, sx, sy, x, y, self.angle]
self.tdw.last_painting_pos = sx, sy
self.last_line_data = last_line
self.adj.set_value(self.slow_tracking)
self.model.brush.reset()
def local_mouse_state(self, last_update=False):
tdw_win = self.tdw.get_window()
display = self.tdw.get_display()
devmgr = display and display.get_device_manager() or None
coredev = devmgr and devmgr.get_client_pointer() or None
if coredev and tdw_win:
win_, x, y, kbmods = tdw_win.get_device_position_double(coredev)
else:
x, y, kbmods = (0.0, 0.0, Gdk.ModifierType(0))
if last_update:
return self.lx, self.ly, kbmods
x, y = self.tdw.display_to_model(x, y)
return x, y, kbmods
def process_line(self):
sx, sy = self.sx, self.sy
x, y, kbmods = self.local_mouse_state(last_update=True)
kbmods ^= self.invert_kbmods # invert using bitwise xor
ctrl = kbmods & Gdk.ModifierType.CONTROL_MASK
shift = kbmods & Gdk.ModifierType.SHIFT_MASK
if self.mode == "CurveLine1":
self.dynamic_curve_1(x, y, sx, sy, self.ex, self.ey)
elif self.mode == "CurveLine2":
ex, ey = self.ex, self.ey
kx, ky = self.kx, self.ky
k2x, k2y = self.k2x, self.k2y
if shift and ctrl:
# moved line end
if not self.flip:
self.dynamic_curve_2(k2x, k2y, x, y, ex, ey, kx, ky)
self.sx, self.sy = x, y
else:
self.dynamic_curve_2(kx, ky, sx, sy, x, y, k2x, k2y)
self.ex, self.ey = x, y
else:
# changed curve shape
self.k2x, self.k2y = x, y
if not self.flip:
self.dynamic_curve_2(x, y, sx, sy, ex, ey, kx, ky)
else:
self.dynamic_curve_2(kx, ky, sx, sy, ex, ey, x, y)
elif self.mode == "EllipseMode":
constrain = False
if ctrl:
x, y = constrain_to_angle(x, y, sx, sy)
constrain = True
if shift:
self.ellipse_rotation_angle(x, y, sx, sy, constrain)
else:
self.ellipse_vec = None
self.dynamic_ellipse(x, y, sx, sy)
else: # if "StraightMode" or "SequenceMode"
if ctrl or shift:
x, y = constrain_to_angle(x, y, sx, sy)
self.dynamic_straight_line(x, y, sx, sy)
return x, y
def ellipse_rotation_angle(self, x, y, sx, sy, constrain):
x1, y1 = normal(sx, sy, x, y)
if self.ellipse_vec is None:
self.ellipse_vec = x1, y1
self.last_angle = self.angle
x2, y2 = self.ellipse_vec
px, py = perpendicular(x2, y2)
pangle = get_angle(x1, y1, px, py)
angle = get_angle(x1, y1, x2, y2)
if pangle > 90.0:
angle = 360 - angle
angle += self.last_angle
if constrain:
angle = constraint_angle(angle)
self.angle = angle
### Line Functions
# Straight Line
def dynamic_straight_line(self, x, y, sx, sy):
self.brush_prep(sx, sy)
entry_p, midpoint_p, junk, prange2, head, tail = self.line_settings()
# Beginning
length, nx, ny = length_and_normal(sx, sy, x, y)
mx, my = multiply_add(sx, sy, nx, ny, 0.25)
self._stroke_to(mx, my, entry_p)
# Middle start
# length = length/2
mx, my = multiply_add(sx, sy, nx, ny, head * length)
self._stroke_to(mx, my, midpoint_p)
# Middle end
mx, my = multiply_add(sx, sy, nx, ny, tail * length)
self._stroke_to(mx, my, midpoint_p)
# End
self._stroke_to(x, y, self.exit_pressure)
# Ellipse
def dynamic_ellipse(self, x, y, sx, sy):
points_in_curve = 360
x1, y1 = difference(sx, sy, x, y)
x1, y1, sin, cos = starting_point_for_ellipse(x1, y1, self.angle)
rx, ry = point_in_ellipse(x1, y1, sin, cos, 0)
self.brush_prep(sx + rx, sy + ry)
entry_p, midpoint_p, prange1, prange2, h, t = self.line_settings()
head = points_in_curve * h
head_range = int(head) + 1
tail = points_in_curve * t
tail_range = int(tail) + 1
tail_length = points_in_curve - tail
# Beginning
px, py = point_in_ellipse(x1, y1, sin, cos, 1)
length, nx, ny = length_and_normal(rx, ry, px, py)
mx, my = multiply_add(rx, ry, nx, ny, 0.25)
self._stroke_to(sx + mx, sy + my, entry_p)
pressure = abs(1 / head * prange1 + entry_p)
self._stroke_to(sx + px, sy + py, pressure)
for degree in xrange(2, head_range):
px, py = point_in_ellipse(x1, y1, sin, cos, degree)
pressure = abs(degree / head * prange1 + entry_p)
self._stroke_to(sx + px, sy + py, pressure)
# Middle
for degree in xrange(head_range, tail_range):
px, py = point_in_ellipse(x1, y1, sin, cos, degree)
self._stroke_to(sx + px, sy + py, midpoint_p)
# End
for degree in xrange(tail_range, points_in_curve + 1):
px, py = point_in_ellipse(x1, y1, sin, cos, degree)
pressure = abs((degree - tail) / tail_length * prange2 + midpoint_p)
self._stroke_to(sx + px, sy + py, pressure)
def dynamic_curve_1(self, cx, cy, sx, sy, ex, ey):
self.brush_prep(sx, sy)
self.draw_curve_1(cx, cy, sx, sy, ex, ey)
def dynamic_curve_2(self, cx, cy, sx, sy, ex, ey, kx, ky):
self.brush_prep(sx, sy)
self.draw_curve_2(cx, cy, sx, sy, ex, ey, kx, ky)
# Curve Straight Line
# Found this page helpful:
# http://www.caffeineowl.com/graphics/2d/vectorial/bezierintro.html
def draw_curve_1(self, cx, cy, sx, sy, ex, ey):
points_in_curve = 100
entry_p, midpoint_p, prange1, prange2, h, t = self.line_settings()
mx, my = midpoint(sx, sy, ex, ey)
length, nx, ny = length_and_normal(mx, my, cx, cy)
cx, cy = multiply_add(mx, my, nx, ny, length * 2)
x1, y1 = difference(sx, sy, cx, cy)
x2, y2 = difference(cx, cy, ex, ey)
head = points_in_curve * h
head_range = int(head) + 1
tail = points_in_curve * t
tail_range = int(tail) + 1
tail_length = points_in_curve - tail
# Beginning
px, py = point_on_curve_1(1, cx, cy, sx, sy, x1, y1, x2, y2)
length, nx, ny = length_and_normal(sx, sy, px, py)
bx, by = multiply_add(sx, sy, nx, ny, 0.25)
self._stroke_to(bx, by, entry_p)
pressure = abs(1 / head * prange1 + entry_p)
self._stroke_to(px, py, pressure)
for i in xrange(2, head_range):
px, py = point_on_curve_1(i, cx, cy, sx, sy, x1, y1, x2, y2)
pressure = abs(i / head * prange1 + entry_p)
self._stroke_to(px, py, pressure)
# Middle
for i in xrange(head_range, tail_range):
px, py = point_on_curve_1(i, cx, cy, sx, sy, x1, y1, x2, y2)
self._stroke_to(px, py, midpoint_p)
# End
for i in xrange(tail_range, points_in_curve + 1):
px, py = point_on_curve_1(i, cx, cy, sx, sy, x1, y1, x2, y2)
pressure = abs((i - tail) / tail_length * prange2 + midpoint_p)
self._stroke_to(px, py, pressure)
def draw_curve_2(self, cx, cy, sx, sy, ex, ey, kx, ky):
points_in_curve = 100
self.brush_prep(sx, sy)
entry_p, midpoint_p, prange1, prange2, h, t = self.line_settings()
mx, my = (cx + sx + ex + kx) / 4.0, (cy + sy + ey + ky) / 4.0
length, nx, ny = length_and_normal(mx, my, cx, cy)
cx, cy = multiply_add(mx, my, nx, ny, length * 2)
length, nx, ny = length_and_normal(mx, my, kx, ky)
kx, ky = multiply_add(mx, my, nx, ny, length * 2)
x1, y1 = difference(sx, sy, cx, cy)
x2, y2 = difference(cx, cy, kx, ky)
x3, y3 = difference(kx, ky, ex, ey)
head = points_in_curve * h
head_range = int(head) + 1
tail = points_in_curve * t
tail_range = int(tail) + 1
tail_length = points_in_curve - tail
# Beginning
px, py = point_on_curve_2(1, cx, cy, sx, sy, kx, ky, x1, y1, x2, y2, x3, y3)
length, nx, ny = length_and_normal(sx, sy, px, py)
bx, by = multiply_add(sx, sy, nx, ny, 0.25)
self._stroke_to(bx, by, entry_p)
pressure = abs(1 / head * prange1 + entry_p)
self._stroke_to(px, py, pressure)
for i in xrange(2, head_range):
px, py = point_on_curve_2(i, cx, cy, sx, sy, kx, ky, x1, y1, x2, y2, x3, y3)
pressure = abs(i / head * prange1 + entry_p)
self._stroke_to(px, py, pressure)
# Middle
for i in xrange(head_range, tail_range):
px, py = point_on_curve_2(i, cx, cy, sx, sy, kx, ky, x1, y1, x2, y2, x3, y3)
self._stroke_to(px, py, midpoint_p)
# End
for i in xrange(tail_range, points_in_curve + 1):
px, py = point_on_curve_2(i, cx, cy, sx, sy, kx, ky, x1, y1, x2, y2, x3, y3)
pressure = abs((i - tail) / tail_length * prange2 + midpoint_p)
self._stroke_to(px, py, pressure)
def _stroke_to(self, x, y, pressure):
# FIXME add control for time, similar to inktool
duration = 1.0
self.stroke_to(
self.model,
duration,
x,
y,
pressure,
0.0,
0.0,
self.doc.tdw.scale,
self.doc.tdw.rotation,
0.0,
auto_split=False,
)
def brush_prep(self, sx, sy):
# Send brush to where the stroke will begin
self.model.brush.reset()
self.brushwork_rollback(self.model)
self.stroke_to(
self.model,
10.0,
sx,
sy,
0.0,
0.0,
0.0,
self.doc.tdw.scale,
self.doc.tdw.rotation,
0.0,
auto_split=False,
)
## Line mode settings
@property
def entry_pressure(self):
adj = self.app.line_mode_settings.adjustments["entry_pressure"]
return adj.get_value()
@property
def midpoint_pressure(self):
adj = self.app.line_mode_settings.adjustments["midpoint_pressure"]
return adj.get_value()
@property
def exit_pressure(self):
adj = self.app.line_mode_settings.adjustments["exit_pressure"]
return adj.get_value()
@property
def head(self):
adj = self.app.line_mode_settings.adjustments["line_head"]
return adj.get_value()
@property
def tail(self):
adj = self.app.line_mode_settings.adjustments["line_tail"]
return adj.get_value()
def line_settings(self):
p1 = self.entry_pressure
p2 = self.midpoint_pressure
p3 = self.exit_pressure
if self.head == 0.0001:
p1 = p2
prange1 = p2 - p1
prange2 = p3 - p2
return p1, p2, prange1, prange2, self.head, self.tail
def redraw_line_cb(self):
# Redraws the line when the line_mode_settings change
last_line = self.last_line_data
if last_line is not None:
current_layer = self.model.layer_stack.current
last_stroke = current_layer.get_last_stroke_info()
if last_line[1] is last_stroke:
# ignore slow_tracking
self.done = True
self.adj = self.app.brush_adjustment["slow_tracking"]
self.slow_tracking = self.adj.get_value()
self.adj.set_value(0)
self.brushwork_rollback(self.model)
self.model.undo()
command = last_line[0]
self.sx, self.sy = last_line[2], last_line[3]
self.ex, self.ey = last_line[4], last_line[5]
x, y = self.ex, self.ey
if command == "EllipseMode":
self.angle = last_line[6]
self.dynamic_ellipse(self.ex, self.ey, self.sx, self.sy)
if command == "CurveLine1":
self.dynamic_straight_line(self.ex, self.ey, self.sx, self.sy)
command = "StraightMode"
if command == "CurveLine2":
x, y = last_line[6], last_line[7]
self.kx, self.ky = last_line[8], last_line[9]
self.k2x, self.k2y = x, y
if (x, y) == (self.kx, self.ky):
self.dynamic_curve_1(x, y, self.sx, self.sy, self.ex, self.ey)
command = "CurveLine1"
else:
self.flip = False
self.dynamic_curve_2(
x, y, self.sx, self.sy, self.ex, self.ey, self.kx, self.ky
)
self.model.sync_pending_changes()
self.record_last_stroke(command, x, y)
class StraightMode(LineModeBase):
ACTION_NAME = "StraightMode"
line_mode = "StraightMode"
@classmethod
def get_name(cls):
return _("Lines and Curves")
def get_usage(self):
return _(
"Draw straight lines; Shift adds curves, "
"Shift + Ctrl moves line ends, "
"Ctrl constrains angle"
)
class SequenceMode(LineModeBase):
ACTION_NAME = "SequenceMode"
line_mode = "SequenceMode"
@classmethod
def get_name(cls):
return _("Connected Lines")
def get_usage(cls):
return _(
"Draw a sequence of lines; Shift adds curves, " "Ctrl constrains angle"
)
class EllipseMode(LineModeBase):
ACTION_NAME = "EllipseMode"
line_mode = "EllipseMode"
@classmethod
def get_name(cls):
return _("Ellipses and Circles")
def get_usage(self):
return _("Draw ellipses; Shift rotates, Ctrl constrains ratio/angle")
## Curve Math
def point_on_curve_1(t, cx, cy, sx, sy, x1, y1, x2, y2):
ratio = t / 100.0
x3, y3 = multiply_add(sx, sy, x1, y1, ratio)
x4, y4 = multiply_add(cx, cy, x2, y2, ratio)
x5, y5 = difference(x3, y3, x4, y4)
x, y = multiply_add(x3, y3, x5, y5, ratio)
return x, y
def point_on_curve_2(t, cx, cy, sx, sy, kx, ky, x1, y1, x2, y2, x3, y3):
ratio = t / 100.0
x4, y4 = multiply_add(sx, sy, x1, y1, ratio)
x5, y5 = multiply_add(cx, cy, x2, y2, ratio)
x6, y6 = multiply_add(kx, ky, x3, y3, ratio)
x1, y1 = difference(x4, y4, x5, y5)
x2, y2 = difference(x5, y5, x6, y6)
x4, y4 = multiply_add(x4, y4, x1, y1, ratio)
x5, y5 = multiply_add(x5, y5, x2, y2, ratio)
x1, y1 = difference(x4, y4, x5, y5)
x, y = multiply_add(x4, y4, x1, y1, ratio)
return x, y
## Ellipse Math
def starting_point_for_ellipse(x, y, rotate):
# Rotate starting point
r = math.radians(rotate)
sin = math.sin(r)
cos = math.cos(r)
x, y = rotate_ellipse(x, y, cos, sin)
return x, y, sin, cos
def point_in_ellipse(x, y, r_sin, r_cos, degree):
# Find point in ellipse
r2 = math.radians(degree)
cos = math.cos(r2)
sin = math.sin(r2)
x = x * cos
y = y * sin
# Rotate Ellipse
x, y = rotate_ellipse(y, x, r_sin, r_cos)
return x, y
def rotate_ellipse(x, y, sin, cos):
x1, y1 = multiply(x, y, sin)
x2, y2 = multiply(x, y, cos)
x = x2 - y1
y = y2 + x1
return x, y
## Vector Math
def get_angle(x1, y1, x2, y2):
dot = dot_product(x1, y1, x2, y2)
if abs(dot) < 1.0:
angle = math.acos(dot) * 180 / math.pi
else:
angle = 0.0
return angle
def constrain_to_angle(x, y, sx, sy):
length, nx, ny = length_and_normal(sx, sy, x, y)
# dot = nx*1 + ny*0 therefore nx
angle = math.acos(nx) * 180 / math.pi
angle = constraint_angle(angle)
ax, ay = angle_normal(ny, angle)
x = sx + ax * length
y = sy + ay * length
return x, y
def constraint_angle(angle):
n = angle // 15
n1 = n * 15
rem = angle - n1
if rem < 7.5:
angle = n * 15.0
else:
angle = (n + 1) * 15.0
return angle
def angle_normal(ny, angle):
if ny < 0.0:
angle = 360.0 - angle
radians = math.radians(angle)
x = math.cos(radians)
y = math.sin(radians)
return x, y
def length_and_normal(x1, y1, x2, y2):
x, y = difference(x1, y1, x2, y2)
length = vector_length(x, y)
if length == 0.0:
x, y = 0.0, 0.0
else:
x, y = x / length, y / length
return length, x, y
def normal(x1, y1, x2, y2):
junk, x, y = length_and_normal(x1, y1, x2, y2)
return x, y
def vector_length(x, y):
length = math.sqrt(x * x + y * y)
return length
def distance(x1, y1, x2, y2):
x, y = difference(x1, y1, x2, y2)
length = vector_length(x, y)
return length
def dot_product(x1, y1, x2, y2):
return x1 * x2 + y1 * y2
def multiply_add(x1, y1, x2, y2, d):
x3, y3 = multiply(x2, y2, d)
x, y = add(x1, y1, x3, y3)
return x, y
def multiply(x, y, d):
# Multiply vector
x = x * d
y = y * d
return x, y
def add(x1, y1, x2, y2):
# Add vectors
x = x1 + x2
y = y1 + y2
return x, y
def difference(x1, y1, x2, y2):
# Difference in x and y between two points
x = x2 - x1
y = y2 - y1
return x, y
def midpoint(x1, y1, x2, y2):
# Midpoint between to points
x = (x1 + x2) / 2.0
y = (y1 + y2) / 2.0
return x, y
def perpendicular(x1, y1):
# Swap x and y, then flip one sign to give vector at 90 degree
x = -y1
y = x1
return x, y
|
term2048 | game | # -*- coding: UTF-8 -*-
"""
Game logic
"""
from __future__ import print_function
import atexit
import math
import os
import os.path
import sys
from colorama import Fore, Style, init
init(autoreset=True)
from term2048 import keypress
from term2048.board import Board
class Game(object):
"""
A 2048 game
"""
__dirs = {
keypress.UP: Board.UP,
keypress.DOWN: Board.DOWN,
keypress.LEFT: Board.LEFT,
keypress.RIGHT: Board.RIGHT,
keypress.SPACE: Board.PAUSE,
}
__is_windows = os.name == "nt"
COLORS = {
2: Fore.GREEN,
4: Fore.BLUE + Style.BRIGHT,
8: Fore.CYAN,
16: Fore.RED,
# Don't use MAGENTA directly; it doesn't display well on Windows.
# see https://github.com/bfontaine/term2048/issues/24
32: Fore.MAGENTA + Style.BRIGHT,
64: Fore.CYAN,
128: Fore.BLUE + Style.BRIGHT,
256: Fore.MAGENTA + Style.BRIGHT,
512: Fore.GREEN,
1024: Fore.RED,
2048: Fore.YELLOW,
# just in case people set an higher goal they still have colors
4096: Fore.RED,
8192: Fore.CYAN,
}
# see Game#adjustColors
# these are color replacements for various modes
__color_modes = {
"dark": {
Fore.BLUE: Fore.WHITE,
Fore.BLUE + Style.BRIGHT: Fore.WHITE,
},
"light": {
Fore.YELLOW: Fore.BLACK,
},
}
SCORES_FILE = "%s/.term2048.scores" % os.path.expanduser("~")
STORE_FILE = "%s/.term2048.store" % os.path.expanduser("~")
def __init__(
self,
scores_file=SCORES_FILE,
colors=None,
store_file=STORE_FILE,
clear_screen=True,
mode=None,
azmode=False,
**kws,
):
"""
Create a new game.
scores_file: file to use for the best score (default
is ~/.term2048.scores)
colors: dictionnary with colors to use for each tile
store_file: file that stores game session's snapshot
mode: color mode. This adjust a few colors and can be 'dark' or
'light'. See the adjustColors functions for more info.
other options are passed to the underlying Board object.
"""
self.board = Board(**kws)
self.score = 0
self.scores_file = scores_file
self.store_file = store_file
self.clear_screen = clear_screen
self.best_score = 0
self.__colors = colors or self.COLORS
self.__azmode = azmode
self.loadBestScore()
self.adjustColors(mode)
def adjustColors(self, mode="dark"):
"""
Change a few colors depending on the mode to use. The default mode
doesn't assume anything and avoid using white & black colors. The dark
mode use white and avoid dark blue while the light mode use black and
avoid yellow, to give a few examples.
"""
rp = Game.__color_modes.get(mode, {})
for k, color in self.__colors.items():
self.__colors[k] = rp.get(color, color)
def loadBestScore(self):
"""
load local best score from the default file
"""
try:
with open(self.scores_file, "r") as f:
self.best_score = int(f.readline(), 10)
except:
return False
return True
def saveBestScore(self):
"""
save current best score in the default file
"""
if self.score > self.best_score:
self.best_score = self.score
try:
with open(self.scores_file, "w") as f:
f.write(str(self.best_score))
except:
return False
return True
def incScore(self, pts):
"""
update the current score by adding it the specified number of points
"""
self.score += pts
if self.score > self.best_score:
self.best_score = self.score
def readMove(self):
"""
read and return a move to pass to a board
"""
k = keypress.getKey()
return Game.__dirs.get(k)
def store(self):
"""
save the current game session's score and data for further use
"""
size = self.board.SIZE
cells = []
for i in range(size):
for j in range(size):
cells.append(str(self.board.getCell(j, i)))
score_str = "%s\n%d" % (" ".join(cells), self.score)
try:
with open(self.store_file, "w") as f:
f.write(score_str)
except:
return False
return True
def restore(self):
"""
restore the saved game score and data
"""
size = self.board.SIZE
try:
with open(self.store_file, "r") as f:
lines = f.readlines()
score_str = lines[0]
self.score = int(lines[1])
except:
return False
score_str_list = score_str.split(" ")
count = 0
for i in range(size):
for j in range(size):
value = score_str_list[count]
self.board.setCell(j, i, int(value))
count += 1
return True
def clearScreen(self):
"""Clear the console"""
if self.clear_screen:
os.system("cls" if self.__is_windows else "clear")
else:
print("\n")
def hideCursor(self):
"""
Hide the cursor. Don't forget to call ``showCursor`` to restore
the normal shell behavior. This is a no-op if ``clear_screen`` is
falsy.
"""
if not self.clear_screen:
return
if not self.__is_windows:
sys.stdout.write("\033[?25l")
def showCursor(self):
"""Show the cursor."""
if not self.__is_windows:
sys.stdout.write("\033[?25h")
def loop(self):
"""
main game loop. returns the final score.
"""
pause_key = self.board.PAUSE
margins = {"left": 4, "top": 4, "bottom": 4}
atexit.register(self.showCursor)
try:
self.hideCursor()
while True:
self.clearScreen()
print(self.__str__(margins=margins))
if self.board.won() or not self.board.canMove():
break
m = self.readMove()
if m == pause_key:
self.saveBestScore()
if self.store():
print(
"Game successfully saved. "
"Resume it with `term2048 --resume`."
)
return self.score
print("An error ocurred while saving your game.")
return None
self.incScore(self.board.move(m))
except KeyboardInterrupt:
self.saveBestScore()
return None
self.saveBestScore()
print("You won!" if self.board.won() else "Game Over")
return self.score
def getCellStr(self, x, y): # TODO: refactor regarding issue #11
"""
return a string representation of the cell located at x,y.
"""
c = self.board.getCell(x, y)
if c == 0:
return "." if self.__azmode else " ."
elif self.__azmode:
az = {}
for i in range(1, int(math.log(self.board.goal(), 2))):
az[2**i] = chr(i + 96)
if c not in az:
return "?"
s = az[c]
elif c == 1024:
s = " 1k"
elif c == 2048:
s = " 2k"
else:
s = "%3d" % c
return self.__colors.get(c, Fore.RESET) + s + Style.RESET_ALL
def boardToString(self, margins=None):
"""
return a string representation of the current board.
"""
if margins is None:
margins = {}
b = self.board
rg = range(b.size())
left = " " * margins.get("left", 0)
s = "\n".join(
[left + " ".join([self.getCellStr(x, y) for x in rg]) for y in rg]
)
return s
def __str__(self, margins=None):
if margins is None:
margins = {}
b = self.boardToString(margins=margins)
top = "\n" * margins.get("top", 0)
bottom = "\n" * margins.get("bottom", 0)
scores = " \tScore: %5d Best: %5d\n" % (self.score, self.best_score)
return top + b.replace("\n", scores, 1) + bottom
|
saveddata | drone | # ===============================================================================
# Copyright (C) 2010 Diego Duclos
#
# This file is part of eos.
#
# eos is free software: you can redistribute it and/or modify
# it under the terms of the GNU Lesser General Public License as published by
# the Free Software Foundation, either version 2 of the License, or
# (at your option) any later version.
#
# eos is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public License
# along with eos. If not, see <http://www.gnu.org/licenses/>.
# ===============================================================================
import datetime
from eos.db import saveddata_meta
from eos.saveddata.drone import Drone
from eos.saveddata.fit import Fit
from eos.saveddata.mutator import MutatorDrone
from sqlalchemy import Boolean, Column, DateTime, Float, ForeignKey, Integer, Table
from sqlalchemy.orm import mapper, relation, synonym
from sqlalchemy.orm.collections import attribute_mapped_collection
drones_table = Table(
"drones",
saveddata_meta,
Column("groupID", Integer, primary_key=True),
Column("fitID", Integer, ForeignKey("fits.ID"), nullable=False, index=True),
Column("itemID", Integer, nullable=False),
Column("baseItemID", Integer, nullable=True),
Column("mutaplasmidID", Integer, nullable=True),
Column("amount", Integer, nullable=False),
Column("amountActive", Integer, nullable=False),
Column("projected", Boolean, default=False),
Column("created", DateTime, nullable=True, default=datetime.datetime.now),
Column("modified", DateTime, nullable=True, onupdate=datetime.datetime.now),
Column("projectionRange", Float, nullable=True),
)
mapper(
Drone,
drones_table,
properties={
"ID": synonym("groupID"),
"owner": relation(Fit),
"mutators": relation(
MutatorDrone,
backref="item",
cascade="all,delete-orphan",
collection_class=attribute_mapped_collection("attrID"),
),
},
)
|
Scripts | dictionary | #!/usr/bin/env python
# This is a doctest
"""
==============================
Using Cocoa collection classes
==============================
Cocoa contains a number of collection classes, including dictionaries and
arrays (like python lists) in mutable and immutable variations. We'll
demonstrate their usage using the ``NSMutableDictonary`` class.
We'll start by importing everything we need::
>>> from Foundation import *
Then create an empty dictionary::
>>> d = NSMutableDictionary.dictionary()
>>> d
{}
>>> isinstance(d, dict)
False
>>> isinstance(d, NSMutableDictionary)
True
You can add a new value using the Objective-C API::
>>> d.setObject_forKey_(42, 'key2')
>>> d
{key2 = 42; }
But can also use the familiar python interface:
>>> d['key1'] = u'hello'
>>> d
{key1 = hello; key2 = 42; }
The same is true for fetching elements::
>>> d['key2']
42
>>> d.objectForKey_('key1')
u'hello'
"""
import doctest
import __main__
doctest.testmod(__main__, verbose=1)
|
code | planet_cache | #!/usr/bin/env python
# -*- coding: UTF-8 -*-
"""Planet cache tool.
"""
__authors__ = [ "Scott James Remnant <scott@netsplit.com>",
"Jeff Waugh <jdub@perkypants.org>" ]
__license__ = "Python"
import os
import sys
import time
import ConfigParser
import dbhash
import planet
def usage():
print "Usage: planet-cache [options] CACHEFILE [ITEMID]..."
print
print "Examine and modify information in the Planet cache."
print
print "Channel Commands:"
print " -C, --channel Display known information on the channel"
print " -L, --list List items in the channel"
print " -K, --keys List all keys found in channel items"
print
print "Item Commands (need ITEMID):"
print " -I, --item Display known information about the item(s)"
print " -H, --hide Mark the item(s) as hidden"
print " -U, --unhide Mark the item(s) as not hidden"
print
print "Other Options:"
print " -h, --help Display this help message and exit"
sys.exit(0)
def usage_error(msg, *args):
print >>sys.stderr, msg, " ".join(args)
print >>sys.stderr, "Perhaps you need --help ?"
sys.exit(1)
def print_keys(item, title):
keys = item.keys()
keys.sort()
key_len = max([ len(k) for k in keys ])
print title + ":"
for key in keys:
if item.key_type(key) == item.DATE:
value = time.strftime(planet.TIMEFMT_ISO, item[key])
else:
value = str(item[key])
print " %-*s %s" % (key_len, key, fit_str(value, 74 - key_len))
def fit_str(string, length):
if len(string) <= length:
return string
else:
return string[:length-4] + " ..."
if __name__ == "__main__":
cache_file = None
want_ids = 0
ids = []
command = None
for arg in sys.argv[1:]:
if arg == "-h" or arg == "--help":
usage()
elif arg == "-C" or arg == "--channel":
if command is not None:
usage_error("Only one command option may be supplied")
command = "channel"
elif arg == "-L" or arg == "--list":
if command is not None:
usage_error("Only one command option may be supplied")
command = "list"
elif arg == "-K" or arg == "--keys":
if command is not None:
usage_error("Only one command option may be supplied")
command = "keys"
elif arg == "-I" or arg == "--item":
if command is not None:
usage_error("Only one command option may be supplied")
command = "item"
want_ids = 1
elif arg == "-H" or arg == "--hide":
if command is not None:
usage_error("Only one command option may be supplied")
command = "hide"
want_ids = 1
elif arg == "-U" or arg == "--unhide":
if command is not None:
usage_error("Only one command option may be supplied")
command = "unhide"
want_ids = 1
elif arg.startswith("-"):
usage_error("Unknown option:", arg)
else:
if cache_file is None:
cache_file = arg
elif want_ids:
ids.append(arg)
else:
usage_error("Unexpected extra argument:", arg)
if cache_file is None:
usage_error("Missing expected cache filename")
elif want_ids and not len(ids):
usage_error("Missing expected entry ids")
# Open the cache file directly to get the URL it represents
try:
db = dbhash.open(cache_file)
url = db["url"]
db.close()
except dbhash.bsddb._db.DBError, e:
print >>sys.stderr, cache_file + ":", e.args[1]
sys.exit(1)
except KeyError:
print >>sys.stderr, cache_file + ": Probably not a cache file"
sys.exit(1)
# Now do it the right way :-)
my_planet = planet.Planet(ConfigParser.ConfigParser())
my_planet.cache_directory = os.path.dirname(cache_file)
channel = planet.Channel(my_planet, url)
for item_id in ids:
if not channel.has_item(item_id):
print >>sys.stderr, item_id + ": Not in channel"
sys.exit(1)
# Do the user's bidding
if command == "channel":
print_keys(channel, "Channel Keys")
elif command == "item":
for item_id in ids:
item = channel.get_item(item_id)
print_keys(item, "Item Keys for %s" % item_id)
elif command == "list":
print "Items in Channel:"
for item in channel.items(hidden=1, sorted=1):
print " " + item.id
print " " + time.strftime(planet.TIMEFMT_ISO, item.date)
if hasattr(item, "title"):
print " " + fit_str(item.title, 70)
if hasattr(item, "hidden"):
print " (hidden)"
elif command == "keys":
keys = {}
for item in channel.items():
for key in item.keys():
keys[key] = 1
keys = keys.keys()
keys.sort()
print "Keys used in Channel:"
for key in keys:
print " " + key
print
print "Use --item to output values of particular items."
elif command == "hide":
for item_id in ids:
item = channel.get_item(item_id)
if hasattr(item, "hidden"):
print item_id + ": Already hidden."
else:
item.hidden = "yes"
channel.cache_write()
print "Done."
elif command == "unhide":
for item_id in ids:
item = channel.get_item(item_id)
if hasattr(item, "hidden"):
del(item.hidden)
else:
print item_id + ": Not hidden."
channel.cache_write()
print "Done."
|
extractors | embed | __all__ = ["embed_download"]
import urllib.parse
from ..common import *
from . import bokecc, iqiyi
from .bilibili import bilibili_download
from .dailymotion import dailymotion_download
from .iqiyi import iqiyi_download_by_vid
from .le import letvcloud_download_by_vu
from .netease import netease_download
from .qq import qq_download_by_vid
from .sina import sina_download_by_vid
from .tudou import tudou_download_by_id
from .vimeo import vimeo_download_by_id
from .youku import youku_download_by_vid
"""
refer to http://open.youku.com/tools
"""
youku_embed_patterns = [
"youku\.com/v_show/id_([a-zA-Z0-9=]+)",
"player\.youku\.com/player\.php/sid/([a-zA-Z0-9=]+)/v\.swf",
"loader\.swf\?VideoIDS=([a-zA-Z0-9=]+)",
"player\.youku\.com/embed/([a-zA-Z0-9=]+)",
"YKU.Player\('[a-zA-Z0-9]+',{ client_id: '[a-zA-Z0-9]+', vid: '([a-zA-Z0-9]+)'",
]
"""
http://www.tudou.com/programs/view/html5embed.action?type=0&code=3LS_URGvl54&lcode=&resourceId=0_06_05_99
"""
tudou_embed_patterns = [
"tudou\.com[a-zA-Z0-9\/\?=\&\.\;]+code=([a-zA-Z0-9_-]+)\&",
'www\.tudou\.com/v/([a-zA-Z0-9_-]+)/[^"]*v\.swf',
]
"""
refer to http://open.tudou.com/wiki/video/info
"""
tudou_api_patterns = []
iqiyi_embed_patterns = [
'player\.video\.qiyi\.com/([^/]+)/[^/]+/[^/]+/[^/]+\.swf[^"]+tvId=(\d+)'
]
netease_embed_patterns = ["(http://\w+\.163\.com/movie/[^'\"]+)"]
vimeo_embed_patters = ["player\.vimeo\.com/video/(\d+)"]
dailymotion_embed_patterns = ["www\.dailymotion\.com/embed/video/(\w+)"]
"""
check the share button on http://www.bilibili.com/video/av5079467/
"""
bilibili_embed_patterns = ["static\.hdslb\.com/miniloader\.swf.*aid=(\d+)"]
"""
http://open.iqiyi.com/lib/player.html
"""
iqiyi_patterns = [
r"(?:\"|\')(https?://dispatcher\.video\.qiyi\.com\/disp\/shareplayer\.swf\?.+?)(?:\"|\')",
r"(?:\"|\')(https?://open\.iqiyi\.com\/developer\/player_js\/coopPlayerIndex\.html\?.+?)(?:\"|\')",
]
bokecc_patterns = [r"bokecc\.com/flash/pocle/player\.swf\?siteid=(.+?)&vid=(.{32})"]
recur_limit = 3
def embed_download(url, output_dir=".", merge=True, info_only=False, **kwargs):
content = get_content(url, headers=fake_headers)
found = False
title = match1(content, "<title>([^<>]+)</title>")
vids = matchall(content, youku_embed_patterns)
for vid in set(vids):
found = True
youku_download_by_vid(
vid,
title=title,
output_dir=output_dir,
merge=merge,
info_only=info_only,
**kwargs,
)
vids = matchall(content, tudou_embed_patterns)
for vid in set(vids):
found = True
tudou_download_by_id(
vid,
title=title,
output_dir=output_dir,
merge=merge,
info_only=info_only,
**kwargs,
)
vids = matchall(content, iqiyi_embed_patterns)
for vid in vids:
found = True
iqiyi_download_by_vid(
(vid[1], vid[0]),
title=title,
output_dir=output_dir,
merge=merge,
info_only=info_only,
**kwargs,
)
urls = matchall(content, netease_embed_patterns)
for url in urls:
found = True
netease_download(
url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs
)
urls = matchall(content, vimeo_embed_patters)
for url in urls:
found = True
vimeo_download_by_id(
url,
title=title,
output_dir=output_dir,
merge=merge,
info_only=info_only,
referer=url,
**kwargs,
)
urls = matchall(content, dailymotion_embed_patterns)
for url in urls:
found = True
dailymotion_download(
url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs
)
aids = matchall(content, bilibili_embed_patterns)
for aid in aids:
found = True
url = "http://www.bilibili.com/video/av%s/" % aid
bilibili_download(
url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs
)
iqiyi_urls = matchall(content, iqiyi_patterns)
for url in iqiyi_urls:
found = True
iqiyi.download(
url, output_dir=output_dir, merge=merge, info_only=info_only, **kwargs
)
bokecc_metas = matchall(content, bokecc_patterns)
for meta in bokecc_metas:
found = True
bokecc.bokecc_download_by_id(
meta[1], output_dir=output_dir, merge=merge, info_only=info_only, **kwargs
)
if found:
return True
# Try harder, check all iframes
if "recur_lv" not in kwargs or kwargs["recur_lv"] < recur_limit:
r = kwargs.get("recur_lv")
if r is None:
r = 1
else:
r += 1
iframes = matchall(content, [r"<iframe.+?src=(?:\"|\')(.*?)(?:\"|\')"])
for iframe in iframes:
if not iframe.startswith("http"):
src = urllib.parse.urljoin(url, iframe)
else:
src = iframe
found = embed_download(
src,
output_dir=output_dir,
merge=merge,
info_only=info_only,
recur_lv=r,
**kwargs,
)
if found:
return True
if not found and "recur_lv" not in kwargs:
raise NotImplementedError(url)
else:
return found
site_info = "any.any"
download = embed_download
download_playlist = playlist_not_supported("any.any")
|
connectors | bookwyrm_connector | """ using another bookwyrm instance as a source of book data """
from __future__ import annotations
from typing import Any, Iterator
from bookwyrm import activitypub, models
from bookwyrm.book_search import SearchResult
from .abstract_connector import AbstractMinimalConnector
class Connector(AbstractMinimalConnector):
"""this is basically just for search"""
def get_or_create_book(self, remote_id: str) -> models.Edition:
return activitypub.resolve_remote_id(remote_id, model=models.Edition)
def parse_search_data(
self, data: list[dict[str, Any]], min_confidence: float
) -> Iterator[SearchResult]:
for search_result in data:
search_result["connector"] = self
yield SearchResult(**search_result)
def parse_isbn_search_data(
self, data: list[dict[str, Any]]
) -> Iterator[SearchResult]:
for search_result in data:
search_result["connector"] = self
yield SearchResult(**search_result)
|
lib | pidfile | import os
import psutil
from sglib.constants import MAJOR_VERSION
from sglib.lib import util
from sglib.log import LOG
def check_pidfile(path):
"""Check if a process is running.
The process must create a pidfile for this to work
"""
if os.path.exists(path):
text = util.read_file_text(path).strip()
try:
pid = int(text)
except Exception as ex:
LOG.exception(ex)
LOG.error(f"Invalid pidfile: {text}")
os.remove(path)
return None
LOG.warning(f"{path} exists with pid {pid}")
if not psutil.pid_exists(pid):
LOG.info(f"pid {pid} no longer exists, deleting pidfile")
os.remove(path)
return None
proc = psutil.Process(pid)
cmdline = " ".join(proc.cmdline())
if MAJOR_VERSION not in cmdline:
LOG.info(f"Another process reclaimed {pid}: {cmdline}")
os.remove(path)
return None
return pid
def create_pidfile(
path,
pid=str(os.getpid()),
):
LOG.info(f"Creating pidfile {path} : {pid}")
util.write_file_text(path, pid)
|
formats | ac3 | # -*- coding: utf-8 -*-
#
# Picard, the next-generation MusicBrainz tagger
#
# Copyright (C) 2019-2020 Philipp Wolfer
# Copyright (C) 2020-2021 Laurent Monin
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License
# as published by the Free Software Foundation; either version 2
# of the License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
import mutagen
from picard import log
from picard.config import get_config
from picard.formats.apev2 import APEv2File
from picard.util import encode_filename
from .mutagenext import ac3
class AC3File(APEv2File):
EXTENSIONS = [".ac3", ".eac3"]
NAME = "AC-3"
_File = ac3.AC3APEv2
def _info(self, metadata, file):
super()._info(metadata, file)
if hasattr(file.info, "codec") and file.info.codec == "ec-3":
format = "Enhanced AC-3"
else:
format = self.NAME
if file.tags:
metadata["~format"] = "%s (APEv2)" % format
else:
metadata["~format"] = format
def _save(self, filename, metadata):
config = get_config()
if config.setting["ac3_save_ape"]:
super()._save(filename, metadata)
elif config.setting["remove_ape_from_ac3"]:
try:
mutagen.apev2.delete(encode_filename(filename))
except BaseException:
log.exception("Error removing APEv2 tags from %s", filename)
@classmethod
def supports_tag(cls, name):
config = get_config()
if config.setting["ac3_save_ape"]:
return APEv2File.supports_tag(name)
else:
return False
|
installer | test_framerates | """
@file
@brief Utility to determine which sample rates divide evenly by which frame rates
@author Jonathan Thomas <jonathan@openshot.org>
@section LICENSE
Copyright (c) 2008-2020 OpenShot Studios, LLC
(http://www.openshotstudios.com). This file is part of
OpenShot Video Editor (http://www.openshot.org), an open-source project
dedicated to delivering high quality video editing and animation solutions
to the world.
OpenShot Video Editor is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
OpenShot Video Editor is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with OpenShot Library. If not, see <http://www.gnu.org/licenses/>.
"""
import math
import os
import openshot
from src.classes import info
# Try to get the security-patched XML functions from defusedxml
try:
from defusedxml import minidom as xml
except ImportError:
from xml.dom import minidom as xml
all_fps = []
all_rates = [
8000.0,
11025.0,
16000.0,
22050.0,
44100.0,
48000.0,
88200.0,
96000.0,
176400.0,
192000.0,
3528000.0,
384000.0,
]
# Loop through all profiles (include any FPS used in OpenShot)
for file in os.listdir(info.PROFILES_PATH):
filepath = os.path.join(info.PROFILES_PATH, file)
profile = openshot.Profile(filepath)
fps_num = profile.info.fps.num
fps_den = profile.info.fps.den
fps = float(fps_num) / float(fps_den)
if fps not in all_fps:
all_fps.append(fps)
# Loop through all export presets (include any sample rates used in OpenShot)
for file in os.listdir(info.EXPORT_PRESETS_PATH):
filepath = os.path.join(info.EXPORT_PRESETS_PATH, file)
xmldoc = xml.parse(filepath)
sampleRate = float(
xmldoc.getElementsByTagName("samplerate")[0].childNodes[0].nodeValue
)
if sampleRate not in all_rates:
all_rates.append(sampleRate)
# Display all common fps and sample rates
print("---------------------------------")
print("Common Sample Rates / Common FPS")
print("---------------------------------")
print("FPS: %s" % sorted(all_fps))
print("Sample Rates: %s" % sorted(all_rates))
print("---------------------------------")
print("GOOD = Sample rate evenly divisible by frame rate")
print("BAD = Sample rate NOT evenly divisible by frame rate")
print("---------------------------------\n")
print("(Sample Rate / FPS) = Samples Per Frame")
print("---------------------------------")
for fps in sorted(all_fps):
for rate in sorted(all_rates):
samples_per_frame = rate / fps
has_remainder = math.fmod(rate, fps)
if not has_remainder:
# Print good case with rounding to 2 digits
print("GOOD:\t%s / %.2f =\t%d" % (rate, fps, samples_per_frame))
else:
# Print bad case, without rounding the samples per frame
print(
"BAD:\t%s / %.2f =\t%s.%s"
% (
rate,
fps,
str(samples_per_frame).split(".")[0],
str(samples_per_frame).split(".")[1][:2],
)
)
|
extractor | hbo | # coding: utf-8
from __future__ import unicode_literals
import re
from ..utils import int_or_none, parse_duration, urljoin, xpath_element, xpath_text
from .common import InfoExtractor
class HBOBaseIE(InfoExtractor):
_FORMATS_INFO = {
"pro7": {
"width": 1280,
"height": 720,
},
"1920": {
"width": 1280,
"height": 720,
},
"pro6": {
"width": 768,
"height": 432,
},
"640": {
"width": 768,
"height": 432,
},
"pro5": {
"width": 640,
"height": 360,
},
"highwifi": {
"width": 640,
"height": 360,
},
"high3g": {
"width": 640,
"height": 360,
},
"medwifi": {
"width": 400,
"height": 224,
},
"med3g": {
"width": 400,
"height": 224,
},
}
def _extract_info(self, url, display_id):
video_data = self._download_xml(url, display_id)
video_id = xpath_text(video_data, "id", fatal=True)
episode_title = title = xpath_text(video_data, "title", fatal=True)
series = xpath_text(video_data, "program")
if series:
title = "%s - %s" % (series, title)
formats = []
for source in xpath_element(video_data, "videos", "sources", True):
if source.tag == "size":
path = xpath_text(source, ".//path")
if not path:
continue
width = source.attrib.get("width")
format_info = self._FORMATS_INFO.get(width, {})
height = format_info.get("height")
fmt = {
"url": path,
"format_id": "http%s" % ("-%dp" % height if height else ""),
"width": format_info.get("width"),
"height": height,
}
rtmp = re.search(
r"^(?P<url>rtmpe?://[^/]+/(?P<app>.+))/(?P<playpath>mp4:.+)$", path
)
if rtmp:
fmt.update(
{
"url": rtmp.group("url"),
"play_path": rtmp.group("playpath"),
"app": rtmp.group("app"),
"ext": "flv",
"format_id": fmt["format_id"].replace("http", "rtmp"),
}
)
formats.append(fmt)
else:
video_url = source.text
if not video_url:
continue
if source.tag == "tarball":
formats.extend(
self._extract_m3u8_formats(
video_url.replace(".tar", "/base_index_w8.m3u8"),
video_id,
"mp4",
"m3u8_native",
m3u8_id="hls",
fatal=False,
)
)
elif source.tag == "hls":
m3u8_formats = self._extract_m3u8_formats(
video_url.replace(".tar", "/base_index.m3u8"),
video_id,
"mp4",
"m3u8_native",
m3u8_id="hls",
fatal=False,
)
for f in m3u8_formats:
if f.get("vcodec") == "none" and not f.get("tbr"):
f["tbr"] = int_or_none(
self._search_regex(
r"-(\d+)k/", f["url"], "tbr", default=None
)
)
formats.extend(m3u8_formats)
elif source.tag == "dash":
formats.extend(
self._extract_mpd_formats(
video_url.replace(".tar", "/manifest.mpd"),
video_id,
mpd_id="dash",
fatal=False,
)
)
else:
format_info = self._FORMATS_INFO.get(source.tag, {})
formats.append(
{
"format_id": "http-%s" % source.tag,
"url": video_url,
"width": format_info.get("width"),
"height": format_info.get("height"),
}
)
self._sort_formats(formats)
thumbnails = []
card_sizes = xpath_element(video_data, "titleCardSizes")
if card_sizes is not None:
for size in card_sizes:
path = xpath_text(size, "path")
if not path:
continue
width = int_or_none(size.get("width"))
thumbnails.append(
{
"id": width,
"url": path,
"width": width,
}
)
subtitles = None
caption_url = xpath_text(video_data, "captionUrl")
if caption_url:
subtitles = {
"en": [{"url": caption_url, "ext": "ttml"}],
}
return {
"id": video_id,
"title": title,
"duration": parse_duration(xpath_text(video_data, "duration/tv14")),
"series": series,
"episode": episode_title,
"formats": formats,
"thumbnails": thumbnails,
"subtitles": subtitles,
}
class HBOIE(HBOBaseIE):
IE_NAME = "hbo"
_VALID_URL = (
r"https?://(?:www\.)?hbo\.com/(?:video|embed)(?:/[^/]+)*/(?P<id>[^/?#]+)"
)
_TEST = {
"url": "https://www.hbo.com/video/game-of-thrones/seasons/season-8/videos/trailer",
"md5": "8126210656f433c452a21367f9ad85b3",
"info_dict": {
"id": "22113301",
"ext": "mp4",
"title": "Game of Thrones - Trailer",
},
"expected_warnings": ["Unknown MIME type application/mp4 in DASH manifest"],
}
def _real_extract(self, url):
display_id = self._match_id(url)
webpage = self._download_webpage(url, display_id)
location_path = self._parse_json(
self._html_search_regex(r'data-state="({.+?})"', webpage, "state"),
display_id,
)["video"]["locationUrl"]
return self._extract_info(urljoin(url, location_path), display_id)
|
puddlestuff | funcprint | # -*- coding: utf-8 -*-
import re
from copy import copy
from functools import partial
from .constants import NO, YES
pattern = re.compile(r"(%\d+\(.+\))|([\\]*\$\d+)")
def perfunc(match, d):
matchtext = match.group()
if matchtext.startswith("\\"):
return matchtext[1:]
try:
number = int(matchtext[1:])
if number >= len(d):
return ""
return d[number]
except ValueError:
text = matchtext[1:-1]
if pattern.search(text):
try:
subfunc = partial(func, d=d)
return pattern.sub(subfunc, text)
except KeyError:
return ""
return matchtext
def func(match, d):
matchtext = match.group()
if matchtext.startswith("\\"):
return matchtext[1:]
try:
number = int(matchtext[1:])
if number >= len(d):
return ""
if d[number] is False:
d[number] = NO
elif d[number] is True:
d[number] = YES
elif isinstance(d[number], int):
d[number] = str(d[number])
elif not isinstance(d[number], str):
if d[number]:
d[number] = YES
else:
d[number] = NO
return d[number]
except ValueError:
number = int(re.search(r"(\d+)", matchtext).group())
if number >= len(d):
return ""
if d[number] is False:
d[number] = YES
elif d[number] is True:
d[number] = NO
elif isinstance(d[number], int):
d[number] = str(d[number])
elif not isinstance(d[number], str):
if d[number]:
d[number] = YES
else:
d[number] = NO
text = re.search(r"%\d+\((.+)\)", matchtext).group(1)
permatch = pattern.search(text)
if permatch:
try:
subfunc = partial(perfunc, d=d)
return pattern.sub(subfunc, text)
except (KeyError, IndexError):
return ""
return text
def pprint(text, args):
args = copy(args)
f = partial(func, d=args)
return pattern.sub(f, text)
|
extractor | theplatform | # coding: utf-8
from __future__ import unicode_literals
import binascii
import hashlib
import hmac
import re
import time
from ..compat import compat_parse_qs, compat_urllib_parse_urlparse
from ..utils import (
ExtractorError,
determine_ext,
find_xpath_attr,
float_or_none,
int_or_none,
mimetype2ext,
sanitized_Request,
unsmuggle_url,
update_url_query,
xpath_with_ns,
)
from .adobepass import AdobePassIE
from .once import OnceIE
default_ns = "http://www.w3.org/2005/SMIL21/Language"
_x = lambda p: xpath_with_ns(p, {"smil": default_ns})
class ThePlatformBaseIE(OnceIE):
_TP_TLD = "com"
def _extract_theplatform_smil(
self, smil_url, video_id, note="Downloading SMIL data"
):
meta = self._download_xml(
smil_url,
video_id,
note=note,
query={"format": "SMIL"},
headers=self.geo_verification_headers(),
)
error_element = find_xpath_attr(meta, _x(".//smil:ref"), "src")
if error_element is not None:
exception = find_xpath_attr(
error_element, _x(".//smil:param"), "name", "exception"
)
if exception is not None:
if exception.get("value") == "GeoLocationBlocked":
self.raise_geo_restricted(error_element.attrib["abstract"])
elif error_element.attrib["src"].startswith(
"http://link.theplatform.%s/s/errorFiles/Unavailable."
% self._TP_TLD
):
raise ExtractorError(
error_element.attrib["abstract"], expected=True
)
smil_formats = self._parse_smil_formats(
meta,
smil_url,
video_id,
namespace=default_ns,
# the parameters are from syfy.com, other sites may use others,
# they also work for nbc.com
f4m_params={"g": "UXWGVKRWHFSP", "hdcore": "3.0.3"},
transform_rtmp_url=lambda streamer, src: (streamer, "mp4:" + src),
)
formats = []
for _format in smil_formats:
if OnceIE.suitable(_format["url"]):
formats.extend(self._extract_once_formats(_format["url"]))
else:
media_url = _format["url"]
if determine_ext(media_url) == "m3u8":
hdnea2 = self._get_cookies(media_url).get("hdnea2")
if hdnea2:
_format["url"] = update_url_query(
media_url, {"hdnea3": hdnea2.value}
)
formats.append(_format)
subtitles = self._parse_smil_subtitles(meta, default_ns)
return formats, subtitles
def _download_theplatform_metadata(self, path, video_id):
info_url = "http://link.theplatform.%s/s/%s?format=preview" % (
self._TP_TLD,
path,
)
return self._download_json(info_url, video_id)
def _parse_theplatform_metadata(self, info):
subtitles = {}
captions = info.get("captions")
if isinstance(captions, list):
for caption in captions:
lang, src, mime = (
caption.get("lang", "en"),
caption.get("src"),
caption.get("type"),
)
subtitles.setdefault(lang, []).append(
{
"ext": mimetype2ext(mime),
"url": src,
}
)
duration = info.get("duration")
tp_chapters = info.get("chapters", [])
chapters = []
if tp_chapters:
def _add_chapter(start_time, end_time):
start_time = float_or_none(start_time, 1000)
end_time = float_or_none(end_time, 1000)
if start_time is None or end_time is None:
return
chapters.append(
{
"start_time": start_time,
"end_time": end_time,
}
)
for chapter in tp_chapters[:-1]:
_add_chapter(chapter.get("startTime"), chapter.get("endTime"))
_add_chapter(
tp_chapters[-1].get("startTime"),
tp_chapters[-1].get("endTime") or duration,
)
return {
"title": info["title"],
"subtitles": subtitles,
"description": info["description"],
"thumbnail": info["defaultThumbnailUrl"],
"duration": float_or_none(duration, 1000),
"timestamp": int_or_none(info.get("pubDate"), 1000) or None,
"uploader": info.get("billingCode"),
"chapters": chapters,
}
def _extract_theplatform_metadata(self, path, video_id):
info = self._download_theplatform_metadata(path, video_id)
return self._parse_theplatform_metadata(info)
class ThePlatformIE(ThePlatformBaseIE, AdobePassIE):
_VALID_URL = r"""(?x)
(?:https?://(?:link|player)\.theplatform\.com/[sp]/(?P<provider_id>[^/]+)/
(?:(?:(?:[^/]+/)+select/)?(?P<media>media/(?:guid/\d+/)?)?|(?P<config>(?:[^/\?]+/(?:swf|config)|onsite)/select/))?
|theplatform:)(?P<id>[^/\?&]+)"""
_TESTS = [
{
# from http://www.metacafe.com/watch/cb-e9I_cZgTgIPd/blackberrys_big_bold_z30/
"url": "http://link.theplatform.com/s/dJ5BDC/e9I_cZgTgIPd/meta.smil?format=smil&Tracking=true&mbr=true",
"info_dict": {
"id": "e9I_cZgTgIPd",
"ext": "flv",
"title": "Blackberry's big, bold Z30",
"description": "The Z30 is Blackberry's biggest, baddest mobile messaging device yet.",
"duration": 247,
"timestamp": 1383239700,
"upload_date": "20131031",
"uploader": "CBSI-NEW",
},
"params": {
# rtmp download
"skip_download": True,
},
"skip": "404 Not Found",
},
{
# from http://www.cnet.com/videos/tesla-model-s-a-second-step-towards-a-cleaner-motoring-future/
"url": "http://link.theplatform.com/s/kYEXFC/22d_qsQ6MIRT",
"info_dict": {
"id": "22d_qsQ6MIRT",
"ext": "flv",
"description": "md5:ac330c9258c04f9d7512cf26b9595409",
"title": "Tesla Model S: A second step towards a cleaner motoring future",
"timestamp": 1426176191,
"upload_date": "20150312",
"uploader": "CBSI-NEW",
},
"params": {
# rtmp download
"skip_download": True,
},
},
{
"url": "https://player.theplatform.com/p/D6x-PC/pulse_preview/embed/select/media/yMBg9E8KFxZD",
"info_dict": {
"id": "yMBg9E8KFxZD",
"ext": "mp4",
"description": "md5:644ad9188d655b742f942bf2e06b002d",
"title": "HIGHLIGHTS: USA bag first ever series Cup win",
"uploader": "EGSM",
},
},
{
"url": "http://player.theplatform.com/p/NnzsPC/widget/select/media/4Y0TlYUr_ZT7",
"only_matching": True,
},
{
"url": "http://player.theplatform.com/p/2E2eJC/nbcNewsOffsite?guid=tdy_or_siri_150701",
"md5": "fb96bb3d85118930a5b055783a3bd992",
"info_dict": {
"id": "tdy_or_siri_150701",
"ext": "mp4",
"title": "iPhone Siri’s sassy response to a math question has people talking",
"description": "md5:a565d1deadd5086f3331d57298ec6333",
"duration": 83.0,
"thumbnail": r"re:^https?://.*\.jpg$",
"timestamp": 1435752600,
"upload_date": "20150701",
"uploader": "NBCU-NEWS",
},
},
{
# From http://www.nbc.com/the-blacklist/video/sir-crispin-crandall/2928790?onid=137781#vc137781=1
# geo-restricted (US), HLS encrypted with AES-128
"url": "http://player.theplatform.com/p/NnzsPC/onsite_universal/select/media/guid/2410887629/2928790?fwsitesection=nbc_the_blacklist_video_library&autoPlay=true&carouselID=137781",
"only_matching": True,
},
]
@classmethod
def _extract_urls(cls, webpage):
m = re.search(
r"""(?x)
<meta\s+
property=(["'])(?:og:video(?::(?:secure_)?url)?|twitter:player)\1\s+
content=(["'])(?P<url>https?://player\.theplatform\.com/p/.+?)\2
""",
webpage,
)
if m:
return [m.group("url")]
# Are whitespaces ignored in URLs?
# https://github.com/ytdl-org/youtube-dl/issues/12044
matches = re.findall(
r'(?s)<(?:iframe|script)[^>]+src=(["\'])((?:https?:)?//player\.theplatform\.com/p/.+?)\1',
webpage,
)
if matches:
return [re.sub(r"\s", "", list(zip(*matches))[1][0])]
@staticmethod
def _sign_url(url, sig_key, sig_secret, life=600, include_qs=False):
flags = "10" if include_qs else "00"
expiration_date = "%x" % (int(time.time()) + life)
def str_to_hex(str):
return binascii.b2a_hex(str.encode("ascii")).decode("ascii")
def hex_to_bytes(hex):
return binascii.a2b_hex(hex.encode("ascii"))
relative_path = re.match(
r"https?://link\.theplatform\.com/s/([^?]+)", url
).group(1)
clear_text = hex_to_bytes(flags + expiration_date + str_to_hex(relative_path))
checksum = hmac.new(
sig_key.encode("ascii"), clear_text, hashlib.sha1
).hexdigest()
sig = flags + expiration_date + checksum + str_to_hex(sig_secret)
return "%s&sig=%s" % (url, sig)
def _real_extract(self, url):
url, smuggled_data = unsmuggle_url(url, {})
self._initialize_geo_bypass(
{
"countries": smuggled_data.get("geo_countries"),
}
)
mobj = re.match(self._VALID_URL, url)
provider_id = mobj.group("provider_id")
video_id = mobj.group("id")
if not provider_id:
provider_id = "dJ5BDC"
path = provider_id + "/"
if mobj.group("media"):
path += mobj.group("media")
path += video_id
qs_dict = compat_parse_qs(compat_urllib_parse_urlparse(url).query)
if "guid" in qs_dict:
webpage = self._download_webpage(url, video_id)
scripts = re.findall(r'<script[^>]+src="([^"]+)"', webpage)
feed_id = None
# feed id usually locates in the last script.
# Seems there's no pattern for the interested script filename, so
# I try one by one
for script in reversed(scripts):
feed_script = self._download_webpage(
self._proto_relative_url(script, "http:"),
video_id,
"Downloading feed script",
)
feed_id = self._search_regex(
r'defaultFeedId\s*:\s*"([^"]+)"',
feed_script,
"default feed id",
default=None,
)
if feed_id is not None:
break
if feed_id is None:
raise ExtractorError("Unable to find feed id")
return self.url_result(
"http://feed.theplatform.com/f/%s/%s?byGuid=%s"
% (provider_id, feed_id, qs_dict["guid"][0])
)
if smuggled_data.get("force_smil_url", False):
smil_url = url
# Explicitly specified SMIL (see https://github.com/ytdl-org/youtube-dl/issues/7385)
elif "/guid/" in url:
headers = {}
source_url = smuggled_data.get("source_url")
if source_url:
headers["Referer"] = source_url
request = sanitized_Request(url, headers=headers)
webpage = self._download_webpage(request, video_id)
smil_url = self._search_regex(
r'<link[^>]+href=(["\'])(?P<url>.+?)\1[^>]+type=["\']application/smil\+xml',
webpage,
"smil url",
group="url",
)
path = self._search_regex(
r"link\.theplatform\.com/s/((?:[^/?#&]+/)+[^/?#&]+)", smil_url, "path"
)
smil_url += "?" if "?" not in smil_url else "&" + "formats=m3u,mpeg4"
elif mobj.group("config"):
config_url = url + "&form=json"
config_url = config_url.replace("swf/", "config/")
config_url = config_url.replace("onsite/", "onsite/config/")
config = self._download_json(config_url, video_id, "Downloading config")
if "releaseUrl" in config:
release_url = config["releaseUrl"]
else:
release_url = "http://link.theplatform.com/s/%s?mbr=true" % path
smil_url = release_url + "&formats=MPEG4&manifest=f4m"
else:
smil_url = "http://link.theplatform.com/s/%s?mbr=true" % path
sig = smuggled_data.get("sig")
if sig:
smil_url = self._sign_url(smil_url, sig["key"], sig["secret"])
formats, subtitles = self._extract_theplatform_smil(smil_url, video_id)
self._sort_formats(formats)
ret = self._extract_theplatform_metadata(path, video_id)
combined_subtitles = self._merge_subtitles(ret.get("subtitles", {}), subtitles)
ret.update(
{
"id": video_id,
"formats": formats,
"subtitles": combined_subtitles,
}
)
return ret
class ThePlatformFeedIE(ThePlatformBaseIE):
_URL_TEMPLATE = "%s//feed.theplatform.com/f/%s/%s?form=json&%s"
_VALID_URL = r"https?://feed\.theplatform\.com/f/(?P<provider_id>[^/]+)/(?P<feed_id>[^?/]+)\?(?:[^&]+&)*(?P<filter>by(?:Gui|I)d=(?P<id>[^&]+))"
_TESTS = [
{
# From http://player.theplatform.com/p/7wvmTC/MSNBCEmbeddedOffSite?guid=n_hardball_5biden_140207
"url": "http://feed.theplatform.com/f/7wvmTC/msnbc_video-p-test?form=json&pretty=true&range=-40&byGuid=n_hardball_5biden_140207",
"md5": "6e32495b5073ab414471b615c5ded394",
"info_dict": {
"id": "n_hardball_5biden_140207",
"ext": "mp4",
"title": "The Biden factor: will Joe run in 2016?",
"description": "Could Vice President Joe Biden be preparing a 2016 campaign? Mark Halperin and Sam Stein weigh in.",
"thumbnail": r"re:^https?://.*\.jpg$",
"upload_date": "20140208",
"timestamp": 1391824260,
"duration": 467.0,
"categories": [
"MSNBC/Issues/Democrats",
"MSNBC/Issues/Elections/Election 2016",
],
"uploader": "NBCU-NEWS",
},
},
{
"url": "http://feed.theplatform.com/f/2E2eJC/nnd_NBCNews?byGuid=nn_netcast_180306.Copy.01",
"only_matching": True,
},
]
def _extract_feed_info(
self,
provider_id,
feed_id,
filter_query,
video_id,
custom_fields=None,
asset_types_query={},
account_id=None,
):
real_url = self._URL_TEMPLATE % (
self.http_scheme(),
provider_id,
feed_id,
filter_query,
)
entry = self._download_json(real_url, video_id)["entries"][0]
main_smil_url = (
"http://link.theplatform.com/s/%s/media/guid/%d/%s"
% (provider_id, account_id, entry["guid"])
if account_id
else entry.get("plmedia$publicUrl")
)
formats = []
subtitles = {}
first_video_id = None
duration = None
asset_types = []
for item in entry["media$content"]:
smil_url = item["plfile$url"]
cur_video_id = ThePlatformIE._match_id(smil_url)
if first_video_id is None:
first_video_id = cur_video_id
duration = float_or_none(item.get("plfile$duration"))
file_asset_types = (
item.get("plfile$assetTypes")
or compat_parse_qs(compat_urllib_parse_urlparse(smil_url).query)[
"assetTypes"
]
)
for asset_type in file_asset_types:
if asset_type in asset_types:
continue
asset_types.append(asset_type)
query = {
"mbr": "true",
"formats": item["plfile$format"],
"assetTypes": asset_type,
}
if asset_type in asset_types_query:
query.update(asset_types_query[asset_type])
cur_formats, cur_subtitles = self._extract_theplatform_smil(
update_url_query(main_smil_url or smil_url, query),
video_id,
"Downloading SMIL data for %s" % asset_type,
)
formats.extend(cur_formats)
subtitles = self._merge_subtitles(subtitles, cur_subtitles)
self._sort_formats(formats)
thumbnails = [
{
"url": thumbnail["plfile$url"],
"width": int_or_none(thumbnail.get("plfile$width")),
"height": int_or_none(thumbnail.get("plfile$height")),
}
for thumbnail in entry.get("media$thumbnails", [])
]
timestamp = int_or_none(entry.get("media$availableDate"), scale=1000)
categories = [item["media$name"] for item in entry.get("media$categories", [])]
ret = self._extract_theplatform_metadata(
"%s/%s" % (provider_id, first_video_id), video_id
)
subtitles = self._merge_subtitles(subtitles, ret["subtitles"])
ret.update(
{
"id": video_id,
"formats": formats,
"subtitles": subtitles,
"thumbnails": thumbnails,
"duration": duration,
"timestamp": timestamp,
"categories": categories,
}
)
if custom_fields:
ret.update(custom_fields(entry))
return ret
def _real_extract(self, url):
mobj = re.match(self._VALID_URL, url)
video_id = mobj.group("id")
provider_id = mobj.group("provider_id")
feed_id = mobj.group("feed_id")
filter_query = mobj.group("filter")
return self._extract_feed_info(provider_id, feed_id, filter_query, video_id)
|
PyObjCTest | test_cfxmlparser | from CoreFoundation import *
from PyObjCTools.TestSupport import *
try:
unicode
except NameError:
unicode = str
try:
long
except NameError:
long = int
class TestXMLParser(TestCase):
# Note: This doesn't actually test the API
def testTypes(self):
self.assertIsCFType(CFXMLParserRef)
def testTypeID(self):
self.assertIsInstance(CFXMLParserGetTypeID(), (int, long))
def testConstants(self):
self.assertEqual(kCFXMLParserValidateDocument, (1 << 0))
self.assertEqual(kCFXMLParserSkipMetaData, (1 << 1))
self.assertEqual(kCFXMLParserReplacePhysicalEntities, (1 << 2))
self.assertEqual(kCFXMLParserSkipWhitespace, (1 << 3))
self.assertEqual(kCFXMLParserResolveExternalEntities, (1 << 4))
self.assertEqual(kCFXMLParserAddImpliedAttributes, (1 << 5))
self.assertEqual(kCFXMLParserAllOptions, 0x00FFFFFF)
self.assertEqual(kCFXMLParserNoOptions, 0)
self.assertEqual(kCFXMLStatusParseNotBegun, -2)
self.assertEqual(kCFXMLStatusParseInProgress, -1)
self.assertEqual(kCFXMLStatusParseSuccessful, 0)
self.assertEqual(kCFXMLErrorUnexpectedEOF, 1)
self.assertEqual(kCFXMLErrorUnknownEncoding, 2)
self.assertEqual(kCFXMLErrorEncodingConversionFailure, 3)
self.assertEqual(kCFXMLErrorMalformedProcessingInstruction, 4)
self.assertEqual(kCFXMLErrorMalformedDTD, 5)
self.assertEqual(kCFXMLErrorMalformedName, 6)
self.assertEqual(kCFXMLErrorMalformedCDSect, 7)
self.assertEqual(kCFXMLErrorMalformedCloseTag, 8)
self.assertEqual(kCFXMLErrorMalformedStartTag, 9)
self.assertEqual(kCFXMLErrorMalformedDocument, 10)
self.assertEqual(kCFXMLErrorElementlessDocument, 11)
self.assertEqual(kCFXMLErrorMalformedComment, 12)
self.assertEqual(kCFXMLErrorMalformedCharacterReference, 13)
self.assertEqual(kCFXMLErrorMalformedParsedCharacterData, 14)
self.assertEqual(kCFXMLErrorNoData, 15)
self.assertIsInstance(kCFXMLTreeErrorDescription, unicode)
self.assertIsInstance(kCFXMLTreeErrorLineNumber, unicode)
self.assertIsInstance(kCFXMLTreeErrorLocation, unicode)
self.assertIsInstance(kCFXMLTreeErrorStatusCode, unicode)
@expectedFailure
def testMissingWrappers(self):
self.fail("Missing manual wrappers for CFXMLParser (low prio)")
if __name__ == "__main__":
main()
|
console | console | #
# Copyright (C) 2008-2009 Ido Abramovich <ido.deluge@gmail.com>
# Copyright (C) 2009 Andrew Resch <andrewresch@gmail.com>
#
# This file is part of Deluge and is licensed under GNU General Public License 3.0, or later, with
# the additional special exception to link portions of this program with the OpenSSL library.
# See LICENSE for more details.
#
import fnmatch
import logging
import os
import sys
import deluge.common
from deluge.argparserbase import ArgParserBase, DelugeTextHelpFormatter
from deluge.ui.ui import UI
log = logging.getLogger(__name__)
#
# Note: Cannot import from console.main here because it imports the twisted reactor.
# Console is imported from console/__init__.py loaded by the script entry points
# defined in setup.py
#
def load_commands(command_dir):
def get_command(name):
command = getattr(
__import__(
"deluge.ui.console.cmdline.commands.%s" % name, {}, {}, ["Command"]
),
"Command",
)()
command._name = name
return command
try:
dir_list = fnmatch.filter(os.listdir(command_dir), "*.py")
except OSError:
return {}
commands = []
for filename in dir_list:
if filename.startswith("_"):
continue
cmd = get_command(os.path.splitext(filename)[0])
for cmd_name in [cmd._name] + cmd.aliases:
commands.append((cmd_name, cmd))
return dict(commands)
class LogStream:
out = sys.stdout
def write(self, data):
self.out.write(data)
def flush(self):
self.out.flush()
class Console(UI):
cmd_description = """Console or command-line user interface"""
def __init__(self, *args, **kwargs):
super().__init__("console", *args, log_stream=LogStream(), **kwargs)
group = self.parser.add_argument_group(
_("Console Options"),
_(
"These daemon connect options will be "
"used for commands, or if console ui autoconnect is enabled."
),
)
group.add_argument(
"-d",
"--daemon",
metavar="<ip_addr>",
dest="daemon_addr",
help=_("Deluge daemon IP address to connect to (default 127.0.0.1)"),
default="127.0.0.1",
)
group.add_argument(
"-p",
"--port",
metavar="<port>",
dest="daemon_port",
type=int,
help=_("Deluge daemon port to connect to (default 58846)"),
default="58846",
)
group.add_argument(
"-U",
"--username",
metavar="<user>",
dest="daemon_user",
help=_("Deluge daemon username to use when connecting"),
)
group.add_argument(
"-P",
"--password",
metavar="<pass>",
dest="daemon_pass",
help=_("Deluge daemon password to use when connecting"),
)
# To properly print help message for the console commands ( e.g. deluge-console info -h),
# we add a subparser for each command which will trigger the help/usage when given
from deluge.ui.console.parser import (
ConsoleCommandParser,
) # import here because (see top)
self.console_parser = ConsoleCommandParser(
parents=[self.parser],
add_help=False,
prog=self.parser.prog,
description="Starts the Deluge console interface",
formatter_class=lambda prog: DelugeTextHelpFormatter(
prog, max_help_position=33, width=90
),
)
self.parser.subparser = self.console_parser
self.console_parser.base_parser = self.parser
subparsers = self.console_parser.add_subparsers(
title=_("Console Commands"),
help=_("Description"),
description=_("The following console commands are available:"),
metavar=_("Command"),
dest="command",
)
from deluge.ui.console import UI_PATH # Must import here
self.console_cmds = load_commands(os.path.join(UI_PATH, "cmdline", "commands"))
for cmd in sorted(self.console_cmds):
self.console_cmds[cmd].add_subparser(subparsers)
def start(self):
if self.ui_args is None:
# Started directly by deluge-console script so must find the UI args manually
options, remaining = ArgParserBase(common_help=False).parse_known_args()
self.ui_args = remaining
i = self.console_parser.find_subcommand(args=self.ui_args)
self.console_parser.subcommand = False
self.parser.subcommand = False if i == -1 else True
super().start(self.console_parser)
from deluge.ui.console.main import ConsoleUI # import here because (see top)
def run(options):
try:
c = ConsoleUI(self.options, self.console_cmds, self.parser.log_stream)
return c.start_ui()
except Exception as ex:
log.exception(ex)
raise
return deluge.common.run_profiled(
run,
self.options,
output_file=self.options.profile,
do_profile=self.options.profile,
)
|
cura | MachineAction | # Copyright (c) 2016 Ultimaker B.V.
# Cura is released under the terms of the LGPLv3 or higher.
import os
from typing import Optional
from PyQt6.QtCore import QObject, QUrl, pyqtProperty, pyqtSignal, pyqtSlot
from UM.Logger import Logger
from UM.PluginObject import PluginObject
from UM.PluginRegistry import PluginRegistry
class MachineAction(QObject, PluginObject):
"""Machine actions are actions that are added to a specific machine type.
Examples of such actions are updating the firmware, connecting with remote devices or doing bed leveling. A
machine action can also have a qml, which should contain a :py:class:`cura.MachineAction.MachineAction` item.
When activated, the item will be displayed in a dialog and this object will be added as "manager" (so all
pyqtSlot() functions can be called by calling manager.func())
"""
def __init__(self, key: str, label: str = "") -> None:
"""Create a new Machine action.
:param key: unique key of the machine action
:param label: Human readable label used to identify the machine action.
"""
super().__init__()
self._key = key
self._label = label
self._qml_url = ""
self._view = None
self._finished = False
self._open_as_dialog = True
self._visible = True
labelChanged = pyqtSignal()
visibilityChanged = pyqtSignal()
onFinished = pyqtSignal()
def getKey(self) -> str:
return self._key
def needsUserInteraction(self) -> bool:
"""Whether this action needs to ask the user anything.
If not, we shouldn't present the user with certain screens which otherwise show up.
:return: Defaults to true to be in line with the old behaviour.
"""
return True
@pyqtProperty(str, notify=labelChanged)
def label(self) -> str:
return self._label
def setLabel(self, label: str) -> None:
if self._label != label:
self._label = label
self.labelChanged.emit()
@pyqtSlot()
def reset(self) -> None:
"""Reset the action to it's default state.
This should not be re-implemented by child classes, instead re-implement _reset.
:py:meth:`cura.MachineAction.MachineAction._reset`
"""
self._finished = False
self._reset()
def _reset(self) -> None:
"""Protected implementation of reset.
See also :py:meth:`cura.MachineAction.MachineAction.reset`
"""
pass
@pyqtSlot()
def execute(self) -> None:
self._execute()
def _execute(self) -> None:
"""Protected implementation of execute."""
pass
@pyqtSlot()
def setFinished(self) -> None:
self._finished = True
self._reset()
self.onFinished.emit()
@pyqtProperty(bool, notify=onFinished)
def finished(self) -> bool:
return self._finished
def _createViewFromQML(self) -> Optional["QObject"]:
"""Protected helper to create a view object based on provided QML."""
plugin_path = PluginRegistry.getInstance().getPluginPath(self.getPluginId())
if plugin_path is None:
Logger.error(
f"Cannot create QML view: cannot find plugin path for plugin {self.getPluginId()}"
)
return None
path = os.path.join(plugin_path, self._qml_url)
from cura.CuraApplication import CuraApplication
view = CuraApplication.getInstance().createQmlComponent(path, {"manager": self})
return view
@pyqtProperty(QUrl, constant=True)
def qmlPath(self) -> "QUrl":
plugin_path = PluginRegistry.getInstance().getPluginPath(self.getPluginId())
if plugin_path is None:
Logger.error(
f"Cannot create QML view: cannot find plugin path for plugin {self.getPluginId()}"
)
return QUrl("")
path = os.path.join(plugin_path, self._qml_url)
return QUrl.fromLocalFile(path)
@pyqtSlot(result=QObject)
def getDisplayItem(self) -> Optional["QObject"]:
return self._createViewFromQML()
@pyqtProperty(bool, constant=True)
def shouldOpenAsDialog(self) -> bool:
"""Whether this action will show a dialog.
If not, the action will directly run the function inside execute().
:return: Defaults to true to be in line with the old behaviour.
"""
return self._open_as_dialog
@pyqtSlot()
def setVisible(self, visible: bool) -> None:
if self._visible != visible:
self._visible = visible
self.visibilityChanged.emit()
@pyqtProperty(bool, notify=visibilityChanged)
def visible(self) -> bool:
"""Whether this action button will be visible.
Example: Show only when isLoggedIn
:return: Defaults to true to be in line with the old behaviour.
"""
return self._visible
|
deluge-extractor | gtkui | #
# Copyright (C) 2009 Andrew Resch <andrewresch@gmail.com>
#
# Basic plugin template created by:
# Copyright (C) 2008 Martijn Voncken <mvoncken@gmail.com>
# Copyright (C) 2007-2009 Andrew Resch <andrewresch@gmail.com>
#
# This file is part of Deluge and is licensed under GNU General Public License 3.0, or later, with
# the additional special exception to link portions of this program with the OpenSSL library.
# See LICENSE for more details.
#
import logging
import gi # isort:skip (Required before Gtk import).
gi.require_version("Gtk", "3.0")
# isort:imports-thirdparty
import deluge.component as component
from deluge.plugins.pluginbase import Gtk3PluginBase
from deluge.ui.client import client
from gi.repository import Gtk
# isort:imports-firstparty
# isort:imports-localfolder
from .common import get_resource
log = logging.getLogger(__name__)
class GtkUI(Gtk3PluginBase):
def enable(self):
self.builder = Gtk.Builder()
self.builder.add_from_file(get_resource("extractor_prefs.ui"))
component.get("Preferences").add_page(
_("Extractor"), self.builder.get_object("extractor_prefs_box")
)
component.get("PluginManager").register_hook(
"on_apply_prefs", self.on_apply_prefs
)
component.get("PluginManager").register_hook(
"on_show_prefs", self.on_show_prefs
)
self.on_show_prefs()
def disable(self):
component.get("Preferences").remove_page(_("Extractor"))
component.get("PluginManager").deregister_hook(
"on_apply_prefs", self.on_apply_prefs
)
component.get("PluginManager").deregister_hook(
"on_show_prefs", self.on_show_prefs
)
del self.builder
def on_apply_prefs(self):
log.debug("applying prefs for Extractor")
if client.is_localhost():
path = self.builder.get_object("folderchooser_path").get_filename()
else:
path = self.builder.get_object("entry_path").get_text()
config = {
"extract_path": path,
"use_name_folder": self.builder.get_object("chk_use_name").get_active(),
}
client.extractor.set_config(config)
def on_show_prefs(self):
if client.is_localhost():
self.builder.get_object("folderchooser_path").show()
self.builder.get_object("entry_path").hide()
else:
self.builder.get_object("folderchooser_path").hide()
self.builder.get_object("entry_path").show()
def on_get_config(config):
if client.is_localhost():
self.builder.get_object("folderchooser_path").set_current_folder(
config["extract_path"]
)
else:
self.builder.get_object("entry_path").set_text(config["extract_path"])
self.builder.get_object("chk_use_name").set_active(
config["use_name_folder"]
)
client.extractor.get_config().addCallback(on_get_config)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.