repository_name
stringlengths
5
67
func_path_in_repository
stringlengths
4
234
func_name
stringlengths
0
314
whole_func_string
stringlengths
52
3.87M
language
stringclasses
6 values
func_code_string
stringlengths
52
3.87M
func_code_tokens
listlengths
15
672k
func_documentation_string
stringlengths
1
47.2k
func_documentation_tokens
listlengths
1
3.92k
split_name
stringclasses
1 value
func_code_url
stringlengths
85
339
canonical-ols/acceptable
acceptable/_service.py
AcceptableAPI.changelog
def changelog(self, api_version, doc): """Add a changelog entry for this api.""" doc = textwrap.dedent(doc).strip() self._changelog[api_version] = doc self._changelog_locations[api_version] = get_callsite_location()
python
def changelog(self, api_version, doc): """Add a changelog entry for this api.""" doc = textwrap.dedent(doc).strip() self._changelog[api_version] = doc self._changelog_locations[api_version] = get_callsite_location()
[ "def", "changelog", "(", "self", ",", "api_version", ",", "doc", ")", ":", "doc", "=", "textwrap", ".", "dedent", "(", "doc", ")", ".", "strip", "(", ")", "self", ".", "_changelog", "[", "api_version", "]", "=", "doc", "self", ".", "_changelog_location...
Add a changelog entry for this api.
[ "Add", "a", "changelog", "entry", "for", "this", "api", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_service.py#L375-L379
elifesciences/elife-tools
elifetools/parseJATS.py
title_prefix
def title_prefix(soup): "titlePrefix for article JSON is only articles with certain display_channel values" prefix = None display_channel_match_list = ['feature article', 'insight', 'editorial'] for d_channel in display_channel(soup): if d_channel.lower() in display_channel_match_list: ...
python
def title_prefix(soup): "titlePrefix for article JSON is only articles with certain display_channel values" prefix = None display_channel_match_list = ['feature article', 'insight', 'editorial'] for d_channel in display_channel(soup): if d_channel.lower() in display_channel_match_list: ...
[ "def", "title_prefix", "(", "soup", ")", ":", "prefix", "=", "None", "display_channel_match_list", "=", "[", "'feature article'", ",", "'insight'", ",", "'editorial'", "]", "for", "d_channel", "in", "display_channel", "(", "soup", ")", ":", "if", "d_channel", ...
titlePrefix for article JSON is only articles with certain display_channel values
[ "titlePrefix", "for", "article", "JSON", "is", "only", "articles", "with", "certain", "display_channel", "values" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L37-L45
elifesciences/elife-tools
elifetools/parseJATS.py
title_prefix_json
def title_prefix_json(soup): "titlePrefix with capitalisation changed" prefix = title_prefix(soup) prefix_rewritten = elifetools.json_rewrite.rewrite_json("title_prefix_json", soup, prefix) return prefix_rewritten
python
def title_prefix_json(soup): "titlePrefix with capitalisation changed" prefix = title_prefix(soup) prefix_rewritten = elifetools.json_rewrite.rewrite_json("title_prefix_json", soup, prefix) return prefix_rewritten
[ "def", "title_prefix_json", "(", "soup", ")", ":", "prefix", "=", "title_prefix", "(", "soup", ")", "prefix_rewritten", "=", "elifetools", ".", "json_rewrite", ".", "rewrite_json", "(", "\"title_prefix_json\"", ",", "soup", ",", "prefix", ")", "return", "prefix_...
titlePrefix with capitalisation changed
[ "titlePrefix", "with", "capitalisation", "changed" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L47-L51
elifesciences/elife-tools
elifetools/parseJATS.py
research_organism
def research_organism(soup): "Find the research-organism from the set of kwd-group tags" if not raw_parser.research_organism_keywords(soup): return [] return list(map(node_text, raw_parser.research_organism_keywords(soup)))
python
def research_organism(soup): "Find the research-organism from the set of kwd-group tags" if not raw_parser.research_organism_keywords(soup): return [] return list(map(node_text, raw_parser.research_organism_keywords(soup)))
[ "def", "research_organism", "(", "soup", ")", ":", "if", "not", "raw_parser", ".", "research_organism_keywords", "(", "soup", ")", ":", "return", "[", "]", "return", "list", "(", "map", "(", "node_text", ",", "raw_parser", ".", "research_organism_keywords", "(...
Find the research-organism from the set of kwd-group tags
[ "Find", "the", "research", "-", "organism", "from", "the", "set", "of", "kwd", "-", "group", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L86-L90
elifesciences/elife-tools
elifetools/parseJATS.py
full_research_organism
def full_research_organism(soup): "research-organism list including inline tags, such as italic" if not raw_parser.research_organism_keywords(soup): return [] return list(map(node_contents_str, raw_parser.research_organism_keywords(soup)))
python
def full_research_organism(soup): "research-organism list including inline tags, such as italic" if not raw_parser.research_organism_keywords(soup): return [] return list(map(node_contents_str, raw_parser.research_organism_keywords(soup)))
[ "def", "full_research_organism", "(", "soup", ")", ":", "if", "not", "raw_parser", ".", "research_organism_keywords", "(", "soup", ")", ":", "return", "[", "]", "return", "list", "(", "map", "(", "node_contents_str", ",", "raw_parser", ".", "research_organism_ke...
research-organism list including inline tags, such as italic
[ "research", "-", "organism", "list", "including", "inline", "tags", "such", "as", "italic" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L92-L96
elifesciences/elife-tools
elifetools/parseJATS.py
keywords
def keywords(soup): """ Find the keywords from the set of kwd-group tags which are typically labelled as the author keywords """ if not raw_parser.author_keywords(soup): return [] return list(map(node_text, raw_parser.author_keywords(soup)))
python
def keywords(soup): """ Find the keywords from the set of kwd-group tags which are typically labelled as the author keywords """ if not raw_parser.author_keywords(soup): return [] return list(map(node_text, raw_parser.author_keywords(soup)))
[ "def", "keywords", "(", "soup", ")", ":", "if", "not", "raw_parser", ".", "author_keywords", "(", "soup", ")", ":", "return", "[", "]", "return", "list", "(", "map", "(", "node_text", ",", "raw_parser", ".", "author_keywords", "(", "soup", ")", ")", ")...
Find the keywords from the set of kwd-group tags which are typically labelled as the author keywords
[ "Find", "the", "keywords", "from", "the", "set", "of", "kwd", "-", "group", "tags", "which", "are", "typically", "labelled", "as", "the", "author", "keywords" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L98-L105
elifesciences/elife-tools
elifetools/parseJATS.py
full_keywords
def full_keywords(soup): "author keywords list including inline tags, such as italic" if not raw_parser.author_keywords(soup): return [] return list(map(node_contents_str, raw_parser.author_keywords(soup)))
python
def full_keywords(soup): "author keywords list including inline tags, such as italic" if not raw_parser.author_keywords(soup): return [] return list(map(node_contents_str, raw_parser.author_keywords(soup)))
[ "def", "full_keywords", "(", "soup", ")", ":", "if", "not", "raw_parser", ".", "author_keywords", "(", "soup", ")", ":", "return", "[", "]", "return", "list", "(", "map", "(", "node_contents_str", ",", "raw_parser", ".", "author_keywords", "(", "soup", ")"...
author keywords list including inline tags, such as italic
[ "author", "keywords", "list", "including", "inline", "tags", "such", "as", "italic" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L107-L111
elifesciences/elife-tools
elifetools/parseJATS.py
version_history
def version_history(soup, html_flag=True): "extract the article version history details" convert = lambda xml_string: xml_to_html(html_flag, xml_string) version_history = [] related_object_tags = raw_parser.related_object(raw_parser.article_meta(soup)) for tag in related_object_tags: article...
python
def version_history(soup, html_flag=True): "extract the article version history details" convert = lambda xml_string: xml_to_html(html_flag, xml_string) version_history = [] related_object_tags = raw_parser.related_object(raw_parser.article_meta(soup)) for tag in related_object_tags: article...
[ "def", "version_history", "(", "soup", ",", "html_flag", "=", "True", ")", ":", "convert", "=", "lambda", "xml_string", ":", "xml_to_html", "(", "html_flag", ",", "xml_string", ")", "version_history", "=", "[", "]", "related_object_tags", "=", "raw_parser", "....
extract the article version history details
[ "extract", "the", "article", "version", "history", "details" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L131-L150
elifesciences/elife-tools
elifetools/parseJATS.py
article_id_list
def article_id_list(soup): """return a list of article-id data""" id_list = [] for article_id_tag in raw_parser.article_id(soup): id_details = OrderedDict() set_if_value(id_details, "type", article_id_tag.get("pub-id-type")) set_if_value(id_details, "value", article_id_tag.text) ...
python
def article_id_list(soup): """return a list of article-id data""" id_list = [] for article_id_tag in raw_parser.article_id(soup): id_details = OrderedDict() set_if_value(id_details, "type", article_id_tag.get("pub-id-type")) set_if_value(id_details, "value", article_id_tag.text) ...
[ "def", "article_id_list", "(", "soup", ")", ":", "id_list", "=", "[", "]", "for", "article_id_tag", "in", "raw_parser", ".", "article_id", "(", "soup", ")", ":", "id_details", "=", "OrderedDict", "(", ")", "set_if_value", "(", "id_details", ",", "\"type\"", ...
return a list of article-id data
[ "return", "a", "list", "of", "article", "-", "id", "data" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L163-L172
elifesciences/elife-tools
elifetools/parseJATS.py
copyright_holder_json
def copyright_holder_json(soup): "for json output add a full stop if ends in et al" holder = None permissions_tag = raw_parser.article_permissions(soup) if permissions_tag: holder = node_text(raw_parser.copyright_holder(permissions_tag)) if holder is not None and holder.endswith('et al'): ...
python
def copyright_holder_json(soup): "for json output add a full stop if ends in et al" holder = None permissions_tag = raw_parser.article_permissions(soup) if permissions_tag: holder = node_text(raw_parser.copyright_holder(permissions_tag)) if holder is not None and holder.endswith('et al'): ...
[ "def", "copyright_holder_json", "(", "soup", ")", ":", "holder", "=", "None", "permissions_tag", "=", "raw_parser", ".", "article_permissions", "(", "soup", ")", "if", "permissions_tag", ":", "holder", "=", "node_text", "(", "raw_parser", ".", "copyright_holder", ...
for json output add a full stop if ends in et al
[ "for", "json", "output", "add", "a", "full", "stop", "if", "ends", "in", "et", "al" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L241-L249
elifesciences/elife-tools
elifetools/parseJATS.py
subject_area
def subject_area(soup): """ Find the subject areas from article-categories subject tags """ subject_area = [] tags = raw_parser.subject_area(soup) for tag in tags: subject_area.append(node_text(tag)) return subject_area
python
def subject_area(soup): """ Find the subject areas from article-categories subject tags """ subject_area = [] tags = raw_parser.subject_area(soup) for tag in tags: subject_area.append(node_text(tag)) return subject_area
[ "def", "subject_area", "(", "soup", ")", ":", "subject_area", "=", "[", "]", "tags", "=", "raw_parser", ".", "subject_area", "(", "soup", ")", "for", "tag", "in", "tags", ":", "subject_area", ".", "append", "(", "node_text", "(", "tag", ")", ")", "retu...
Find the subject areas from article-categories subject tags
[ "Find", "the", "subject", "areas", "from", "article", "-", "categories", "subject", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L298-L308
elifesciences/elife-tools
elifetools/parseJATS.py
display_channel
def display_channel(soup): """ Find the subject areas of type display-channel """ display_channel = [] tags = raw_parser.display_channel(soup) for tag in tags: display_channel.append(node_text(tag)) return display_channel
python
def display_channel(soup): """ Find the subject areas of type display-channel """ display_channel = [] tags = raw_parser.display_channel(soup) for tag in tags: display_channel.append(node_text(tag)) return display_channel
[ "def", "display_channel", "(", "soup", ")", ":", "display_channel", "=", "[", "]", "tags", "=", "raw_parser", ".", "display_channel", "(", "soup", ")", "for", "tag", "in", "tags", ":", "display_channel", ".", "append", "(", "node_text", "(", "tag", ")", ...
Find the subject areas of type display-channel
[ "Find", "the", "subject", "areas", "of", "type", "display", "-", "channel" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L321-L331
elifesciences/elife-tools
elifetools/parseJATS.py
category
def category(soup): """ Find the category from subject areas """ category = [] tags = raw_parser.category(soup) for tag in tags: category.append(node_text(tag)) return category
python
def category(soup): """ Find the category from subject areas """ category = [] tags = raw_parser.category(soup) for tag in tags: category.append(node_text(tag)) return category
[ "def", "category", "(", "soup", ")", ":", "category", "=", "[", "]", "tags", "=", "raw_parser", ".", "category", "(", "soup", ")", "for", "tag", "in", "tags", ":", "category", ".", "append", "(", "node_text", "(", "tag", ")", ")", "return", "category...
Find the category from subject areas
[ "Find", "the", "category", "from", "subject", "areas" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L333-L343
elifesciences/elife-tools
elifetools/parseJATS.py
ymd
def ymd(soup): """ Get the year, month and day from child tags """ day = node_text(raw_parser.day(soup)) month = node_text(raw_parser.month(soup)) year = node_text(raw_parser.year(soup)) return (day, month, year)
python
def ymd(soup): """ Get the year, month and day from child tags """ day = node_text(raw_parser.day(soup)) month = node_text(raw_parser.month(soup)) year = node_text(raw_parser.year(soup)) return (day, month, year)
[ "def", "ymd", "(", "soup", ")", ":", "day", "=", "node_text", "(", "raw_parser", ".", "day", "(", "soup", ")", ")", "month", "=", "node_text", "(", "raw_parser", ".", "month", "(", "soup", ")", ")", "year", "=", "node_text", "(", "raw_parser", ".", ...
Get the year, month and day from child tags
[ "Get", "the", "year", "month", "and", "day", "from", "child", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L345-L352
elifesciences/elife-tools
elifetools/parseJATS.py
pub_date
def pub_date(soup): """ Return the publishing date in struct format pub_date_date, pub_date_day, pub_date_month, pub_date_year, pub_date_timestamp Default date_type is pub """ pub_date = first(raw_parser.pub_date(soup, date_type="pub")) if pub_date is None: pub_date = first(raw_parse...
python
def pub_date(soup): """ Return the publishing date in struct format pub_date_date, pub_date_day, pub_date_month, pub_date_year, pub_date_timestamp Default date_type is pub """ pub_date = first(raw_parser.pub_date(soup, date_type="pub")) if pub_date is None: pub_date = first(raw_parse...
[ "def", "pub_date", "(", "soup", ")", ":", "pub_date", "=", "first", "(", "raw_parser", ".", "pub_date", "(", "soup", ",", "date_type", "=", "\"pub\"", ")", ")", "if", "pub_date", "is", "None", ":", "pub_date", "=", "first", "(", "raw_parser", ".", "pub...
Return the publishing date in struct format pub_date_date, pub_date_day, pub_date_month, pub_date_year, pub_date_timestamp Default date_type is pub
[ "Return", "the", "publishing", "date", "in", "struct", "format", "pub_date_date", "pub_date_day", "pub_date_month", "pub_date_year", "pub_date_timestamp", "Default", "date_type", "is", "pub" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L354-L366
elifesciences/elife-tools
elifetools/parseJATS.py
pub_dates
def pub_dates(soup): """ return a list of all the pub dates """ pub_dates = [] tags = raw_parser.pub_date(soup) for tag in tags: pub_date = OrderedDict() copy_attribute(tag.attrs, 'publication-format', pub_date) copy_attribute(tag.attrs, 'date-type', pub_date) cop...
python
def pub_dates(soup): """ return a list of all the pub dates """ pub_dates = [] tags = raw_parser.pub_date(soup) for tag in tags: pub_date = OrderedDict() copy_attribute(tag.attrs, 'publication-format', pub_date) copy_attribute(tag.attrs, 'date-type', pub_date) cop...
[ "def", "pub_dates", "(", "soup", ")", ":", "pub_dates", "=", "[", "]", "tags", "=", "raw_parser", ".", "pub_date", "(", "soup", ")", "for", "tag", "in", "tags", ":", "pub_date", "=", "OrderedDict", "(", ")", "copy_attribute", "(", "tag", ".", "attrs", ...
return a list of all the pub dates
[ "return", "a", "list", "of", "all", "the", "pub", "dates" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L368-L387
elifesciences/elife-tools
elifetools/parseJATS.py
history_date
def history_date(soup, date_type = None): """ Find a date in the history tag for the specific date_type typical date_type values: received, accepted """ if(date_type == None): return None history_date = raw_parser.history_date(soup, date_type) if history_date is None: return...
python
def history_date(soup, date_type = None): """ Find a date in the history tag for the specific date_type typical date_type values: received, accepted """ if(date_type == None): return None history_date = raw_parser.history_date(soup, date_type) if history_date is None: return...
[ "def", "history_date", "(", "soup", ",", "date_type", "=", "None", ")", ":", "if", "(", "date_type", "==", "None", ")", ":", "return", "None", "history_date", "=", "raw_parser", ".", "history_date", "(", "soup", ",", "date_type", ")", "if", "history_date",...
Find a date in the history tag for the specific date_type typical date_type values: received, accepted
[ "Find", "a", "date", "in", "the", "history", "tag", "for", "the", "specific", "date_type", "typical", "date_type", "values", ":", "received", "accepted" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L389-L401
elifesciences/elife-tools
elifetools/parseJATS.py
collection_year
def collection_year(soup): """ Pub date of type collection will hold a year element for VOR articles """ pub_date = first(raw_parser.pub_date(soup, pub_type="collection")) if not pub_date: pub_date = first(raw_parser.pub_date(soup, date_type="collection")) if not pub_date: return...
python
def collection_year(soup): """ Pub date of type collection will hold a year element for VOR articles """ pub_date = first(raw_parser.pub_date(soup, pub_type="collection")) if not pub_date: pub_date = first(raw_parser.pub_date(soup, date_type="collection")) if not pub_date: return...
[ "def", "collection_year", "(", "soup", ")", ":", "pub_date", "=", "first", "(", "raw_parser", ".", "pub_date", "(", "soup", ",", "pub_type", "=", "\"collection\"", ")", ")", "if", "not", "pub_date", ":", "pub_date", "=", "first", "(", "raw_parser", ".", ...
Pub date of type collection will hold a year element for VOR articles
[ "Pub", "date", "of", "type", "collection", "will", "hold", "a", "year", "element", "for", "VOR", "articles" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L493-L508
elifesciences/elife-tools
elifetools/parseJATS.py
abstracts
def abstracts(soup): """ Find the article abstract and format it """ abstracts = [] abstract_tags = raw_parser.abstract(soup) for tag in abstract_tags: abstract = {} abstract["abstract_type"] = tag.get("abstract-type") title_tag = raw_parser.title(tag) if titl...
python
def abstracts(soup): """ Find the article abstract and format it """ abstracts = [] abstract_tags = raw_parser.abstract(soup) for tag in abstract_tags: abstract = {} abstract["abstract_type"] = tag.get("abstract-type") title_tag = raw_parser.title(tag) if titl...
[ "def", "abstracts", "(", "soup", ")", ":", "abstracts", "=", "[", "]", "abstract_tags", "=", "raw_parser", ".", "abstract", "(", "soup", ")", "for", "tag", "in", "abstract_tags", ":", "abstract", "=", "{", "}", "abstract", "[", "\"abstract_type\"", "]", ...
Find the article abstract and format it
[ "Find", "the", "article", "abstract", "and", "format", "it" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L520-L557
elifesciences/elife-tools
elifetools/parseJATS.py
component_doi
def component_doi(soup): """ Look for all object-id of pub-type-id = doi, these are the component DOI tags """ component_doi = [] object_id_tags = raw_parser.object_id(soup, pub_id_type = "doi") # Get components too for later component_list = components(soup) position = 1 for tag...
python
def component_doi(soup): """ Look for all object-id of pub-type-id = doi, these are the component DOI tags """ component_doi = [] object_id_tags = raw_parser.object_id(soup, pub_id_type = "doi") # Get components too for later component_list = components(soup) position = 1 for tag...
[ "def", "component_doi", "(", "soup", ")", ":", "component_doi", "=", "[", "]", "object_id_tags", "=", "raw_parser", ".", "object_id", "(", "soup", ",", "pub_id_type", "=", "\"doi\"", ")", "# Get components too for later", "component_list", "=", "components", "(", ...
Look for all object-id of pub-type-id = doi, these are the component DOI tags
[ "Look", "for", "all", "object", "-", "id", "of", "pub", "-", "type", "-", "id", "=", "doi", "these", "are", "the", "component", "DOI", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L648-L675
elifesciences/elife-tools
elifetools/parseJATS.py
tag_details
def tag_details(tag, nodenames): """ Used in media and graphics to extract data from their parent tags """ details = {} details['type'] = tag.name details['ordinal'] = tag_ordinal(tag) # Ordinal value if tag_details_sibling_ordinal(tag): details['sibling_ordinal'] = tag_details...
python
def tag_details(tag, nodenames): """ Used in media and graphics to extract data from their parent tags """ details = {} details['type'] = tag.name details['ordinal'] = tag_ordinal(tag) # Ordinal value if tag_details_sibling_ordinal(tag): details['sibling_ordinal'] = tag_details...
[ "def", "tag_details", "(", "tag", ",", "nodenames", ")", ":", "details", "=", "{", "}", "details", "[", "'type'", "]", "=", "tag", ".", "name", "details", "[", "'ordinal'", "]", "=", "tag_ordinal", "(", "tag", ")", "# Ordinal value", "if", "tag_details_s...
Used in media and graphics to extract data from their parent tags
[ "Used", "in", "media", "and", "graphics", "to", "extract", "data", "from", "their", "parent", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L726-L747
elifesciences/elife-tools
elifetools/parseJATS.py
media
def media(soup): """ All media tags and some associated data about the related component doi and the parent of that doi (not always present) """ media = [] media_tags = raw_parser.media(soup) position = 1 for tag in media_tags: media_item = {} copy_attribute(tag.attrs...
python
def media(soup): """ All media tags and some associated data about the related component doi and the parent of that doi (not always present) """ media = [] media_tags = raw_parser.media(soup) position = 1 for tag in media_tags: media_item = {} copy_attribute(tag.attrs...
[ "def", "media", "(", "soup", ")", ":", "media", "=", "[", "]", "media_tags", "=", "raw_parser", ".", "media", "(", "soup", ")", "position", "=", "1", "for", "tag", "in", "media_tags", ":", "media_item", "=", "{", "}", "copy_attribute", "(", "tag", "....
All media tags and some associated data about the related component doi and the parent of that doi (not always present)
[ "All", "media", "tags", "and", "some", "associated", "data", "about", "the", "related", "component", "doi", "and", "the", "parent", "of", "that", "doi", "(", "not", "always", "present", ")" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L750-L821
elifesciences/elife-tools
elifetools/parseJATS.py
graphics
def graphics(soup): """ All graphic tags and some associated data about the related component doi and the parent of that doi (not always present), and whether it is part of a figure supplement """ graphics = [] graphic_tags = raw_parser.graphic(soup) position = 1 for tag in graphi...
python
def graphics(soup): """ All graphic tags and some associated data about the related component doi and the parent of that doi (not always present), and whether it is part of a figure supplement """ graphics = [] graphic_tags = raw_parser.graphic(soup) position = 1 for tag in graphi...
[ "def", "graphics", "(", "soup", ")", ":", "graphics", "=", "[", "]", "graphic_tags", "=", "raw_parser", ".", "graphic", "(", "soup", ")", "position", "=", "1", "for", "tag", "in", "graphic_tags", ":", "graphic_item", "=", "{", "}", "copy_attribute", "(",...
All graphic tags and some associated data about the related component doi and the parent of that doi (not always present), and whether it is part of a figure supplement
[ "All", "graphic", "tags", "and", "some", "associated", "data", "about", "the", "related", "component", "doi", "and", "the", "parent", "of", "that", "doi", "(", "not", "always", "present", ")", "and", "whether", "it", "is", "part", "of", "a", "figure", "s...
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L824-L877
elifesciences/elife-tools
elifetools/parseJATS.py
inline_graphics
def inline_graphics(soup): """ inline-graphic tags """ inline_graphics = [] inline_graphic_tags = raw_parser.inline_graphic(soup) position = 1 for tag in inline_graphic_tags: item = {} copy_attribute(tag.attrs, 'xlink:href', item, 'xlink_href') # Get the tag type...
python
def inline_graphics(soup): """ inline-graphic tags """ inline_graphics = [] inline_graphic_tags = raw_parser.inline_graphic(soup) position = 1 for tag in inline_graphic_tags: item = {} copy_attribute(tag.attrs, 'xlink:href', item, 'xlink_href') # Get the tag type...
[ "def", "inline_graphics", "(", "soup", ")", ":", "inline_graphics", "=", "[", "]", "inline_graphic_tags", "=", "raw_parser", ".", "inline_graphic", "(", "soup", ")", "position", "=", "1", "for", "tag", "in", "inline_graphic_tags", ":", "item", "=", "{", "}",...
inline-graphic tags
[ "inline", "-", "graphic", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L879-L906
elifesciences/elife-tools
elifetools/parseJATS.py
self_uri
def self_uri(soup): """ self-uri tags """ self_uri = [] self_uri_tags = raw_parser.self_uri(soup) position = 1 for tag in self_uri_tags: item = {} copy_attribute(tag.attrs, 'xlink:href', item, 'xlink_href') copy_attribute(tag.attrs, 'content-type', item) # ...
python
def self_uri(soup): """ self-uri tags """ self_uri = [] self_uri_tags = raw_parser.self_uri(soup) position = 1 for tag in self_uri_tags: item = {} copy_attribute(tag.attrs, 'xlink:href', item, 'xlink_href') copy_attribute(tag.attrs, 'content-type', item) # ...
[ "def", "self_uri", "(", "soup", ")", ":", "self_uri", "=", "[", "]", "self_uri_tags", "=", "raw_parser", ".", "self_uri", "(", "soup", ")", "position", "=", "1", "for", "tag", "in", "self_uri_tags", ":", "item", "=", "{", "}", "copy_attribute", "(", "t...
self-uri tags
[ "self", "-", "uri", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L908-L934
elifesciences/elife-tools
elifetools/parseJATS.py
supplementary_material
def supplementary_material(soup): """ supplementary-material tags """ supplementary_material = [] supplementary_material_tags = raw_parser.supplementary_material(soup) position = 1 for tag in supplementary_material_tags: item = {} copy_attribute(tag.attrs, 'id', item) ...
python
def supplementary_material(soup): """ supplementary-material tags """ supplementary_material = [] supplementary_material_tags = raw_parser.supplementary_material(soup) position = 1 for tag in supplementary_material_tags: item = {} copy_attribute(tag.attrs, 'id', item) ...
[ "def", "supplementary_material", "(", "soup", ")", ":", "supplementary_material", "=", "[", "]", "supplementary_material_tags", "=", "raw_parser", ".", "supplementary_material", "(", "soup", ")", "position", "=", "1", "for", "tag", "in", "supplementary_material_tags",...
supplementary-material tags
[ "supplementary", "-", "material", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L936-L970
elifesciences/elife-tools
elifetools/parseJATS.py
contrib_email
def contrib_email(contrib_tag): """ Given a contrib tag, look for an email tag, and only return the value if it is not inside an aff tag """ email = [] for email_tag in extract_nodes(contrib_tag, "email"): if email_tag.parent.name != "aff": email.append(email_tag.text) re...
python
def contrib_email(contrib_tag): """ Given a contrib tag, look for an email tag, and only return the value if it is not inside an aff tag """ email = [] for email_tag in extract_nodes(contrib_tag, "email"): if email_tag.parent.name != "aff": email.append(email_tag.text) re...
[ "def", "contrib_email", "(", "contrib_tag", ")", ":", "email", "=", "[", "]", "for", "email_tag", "in", "extract_nodes", "(", "contrib_tag", ",", "\"email\"", ")", ":", "if", "email_tag", ".", "parent", ".", "name", "!=", "\"aff\"", ":", "email", ".", "a...
Given a contrib tag, look for an email tag, and only return the value if it is not inside an aff tag
[ "Given", "a", "contrib", "tag", "look", "for", "an", "email", "tag", "and", "only", "return", "the", "value", "if", "it", "is", "not", "inside", "an", "aff", "tag" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L979-L988
elifesciences/elife-tools
elifetools/parseJATS.py
contrib_phone
def contrib_phone(contrib_tag): """ Given a contrib tag, look for an phone tag """ phone = None if raw_parser.phone(contrib_tag): phone = first(raw_parser.phone(contrib_tag)).text return phone
python
def contrib_phone(contrib_tag): """ Given a contrib tag, look for an phone tag """ phone = None if raw_parser.phone(contrib_tag): phone = first(raw_parser.phone(contrib_tag)).text return phone
[ "def", "contrib_phone", "(", "contrib_tag", ")", ":", "phone", "=", "None", "if", "raw_parser", ".", "phone", "(", "contrib_tag", ")", ":", "phone", "=", "first", "(", "raw_parser", ".", "phone", "(", "contrib_tag", ")", ")", ".", "text", "return", "phon...
Given a contrib tag, look for an phone tag
[ "Given", "a", "contrib", "tag", "look", "for", "an", "phone", "tag" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L990-L997
elifesciences/elife-tools
elifetools/parseJATS.py
contrib_inline_aff
def contrib_inline_aff(contrib_tag): """ Given a contrib tag, look for an aff tag directly inside it """ aff_tags = [] for child_tag in contrib_tag: if child_tag and child_tag.name and child_tag.name == "aff": aff_tags.append(child_tag) return aff_tags
python
def contrib_inline_aff(contrib_tag): """ Given a contrib tag, look for an aff tag directly inside it """ aff_tags = [] for child_tag in contrib_tag: if child_tag and child_tag.name and child_tag.name == "aff": aff_tags.append(child_tag) return aff_tags
[ "def", "contrib_inline_aff", "(", "contrib_tag", ")", ":", "aff_tags", "=", "[", "]", "for", "child_tag", "in", "contrib_tag", ":", "if", "child_tag", "and", "child_tag", ".", "name", "and", "child_tag", ".", "name", "==", "\"aff\"", ":", "aff_tags", ".", ...
Given a contrib tag, look for an aff tag directly inside it
[ "Given", "a", "contrib", "tag", "look", "for", "an", "aff", "tag", "directly", "inside", "it" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L999-L1007
elifesciences/elife-tools
elifetools/parseJATS.py
contrib_xref
def contrib_xref(contrib_tag, ref_type): """ Given a contrib tag, look for an xref tag of type ref_type directly inside the contrib tag """ aff_tags = [] for child_tag in contrib_tag: if (child_tag and child_tag.name and child_tag.name == "xref" and child_tag.get('ref-type') and ...
python
def contrib_xref(contrib_tag, ref_type): """ Given a contrib tag, look for an xref tag of type ref_type directly inside the contrib tag """ aff_tags = [] for child_tag in contrib_tag: if (child_tag and child_tag.name and child_tag.name == "xref" and child_tag.get('ref-type') and ...
[ "def", "contrib_xref", "(", "contrib_tag", ",", "ref_type", ")", ":", "aff_tags", "=", "[", "]", "for", "child_tag", "in", "contrib_tag", ":", "if", "(", "child_tag", "and", "child_tag", ".", "name", "and", "child_tag", ".", "name", "==", "\"xref\"", "and"...
Given a contrib tag, look for an xref tag of type ref_type directly inside the contrib tag
[ "Given", "a", "contrib", "tag", "look", "for", "an", "xref", "tag", "of", "type", "ref_type", "directly", "inside", "the", "contrib", "tag" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1009-L1018
elifesciences/elife-tools
elifetools/parseJATS.py
all_contributors
def all_contributors(soup, detail="brief"): "find all contributors not contrained to only the ones in article meta" contrib_tags = raw_parser.contributors(soup) contributors = format_authors(soup, contrib_tags, detail) return contributors
python
def all_contributors(soup, detail="brief"): "find all contributors not contrained to only the ones in article meta" contrib_tags = raw_parser.contributors(soup) contributors = format_authors(soup, contrib_tags, detail) return contributors
[ "def", "all_contributors", "(", "soup", ",", "detail", "=", "\"brief\"", ")", ":", "contrib_tags", "=", "raw_parser", ".", "contributors", "(", "soup", ")", "contributors", "=", "format_authors", "(", "soup", ",", "contrib_tags", ",", "detail", ")", "return", ...
find all contributors not contrained to only the ones in article meta
[ "find", "all", "contributors", "not", "contrained", "to", "only", "the", "ones", "in", "article", "meta" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1205-L1209
elifesciences/elife-tools
elifetools/parseJATS.py
authors_non_byline
def authors_non_byline(soup, detail="full"): """Non-byline authors for group author members""" # Get a filtered list of contributors, in order to get their group-author-id contrib_type = "author non-byline" contributors_ = contributors(soup, detail) non_byline_authors = [author for author in contrib...
python
def authors_non_byline(soup, detail="full"): """Non-byline authors for group author members""" # Get a filtered list of contributors, in order to get their group-author-id contrib_type = "author non-byline" contributors_ = contributors(soup, detail) non_byline_authors = [author for author in contrib...
[ "def", "authors_non_byline", "(", "soup", ",", "detail", "=", "\"full\"", ")", ":", "# Get a filtered list of contributors, in order to get their group-author-id", "contrib_type", "=", "\"author non-byline\"", "contributors_", "=", "contributors", "(", "soup", ",", "detail", ...
Non-byline authors for group author members
[ "Non", "-", "byline", "authors", "for", "group", "author", "members" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1222-L1234
elifesciences/elife-tools
elifetools/parseJATS.py
refs
def refs(soup): """Find and return all the references""" tags = raw_parser.ref_list(soup) refs = [] position = 1 article_doi = doi(soup) for tag in tags: ref = {} ref['ref'] = ref_text(tag) # ref_id copy_attribute(tag.attrs, "id", ref) # article_title...
python
def refs(soup): """Find and return all the references""" tags = raw_parser.ref_list(soup) refs = [] position = 1 article_doi = doi(soup) for tag in tags: ref = {} ref['ref'] = ref_text(tag) # ref_id copy_attribute(tag.attrs, "id", ref) # article_title...
[ "def", "refs", "(", "soup", ")", ":", "tags", "=", "raw_parser", ".", "ref_list", "(", "soup", ")", "refs", "=", "[", "]", "position", "=", "1", "article_doi", "=", "doi", "(", "soup", ")", "for", "tag", "in", "tags", ":", "ref", "=", "{", "}", ...
Find and return all the references
[ "Find", "and", "return", "all", "the", "references" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1340-L1484
elifesciences/elife-tools
elifetools/parseJATS.py
extract_component_doi
def extract_component_doi(tag, nodenames): """ Used to get component DOI from a tag and confirm it is actually for that tag and it is not for one of its children in the list of nodenames """ component_doi = None if(tag.name == "sub-article"): component_doi = doi_uri_to_doi(node_text(fir...
python
def extract_component_doi(tag, nodenames): """ Used to get component DOI from a tag and confirm it is actually for that tag and it is not for one of its children in the list of nodenames """ component_doi = None if(tag.name == "sub-article"): component_doi = doi_uri_to_doi(node_text(fir...
[ "def", "extract_component_doi", "(", "tag", ",", "nodenames", ")", ":", "component_doi", "=", "None", "if", "(", "tag", ".", "name", "==", "\"sub-article\"", ")", ":", "component_doi", "=", "doi_uri_to_doi", "(", "node_text", "(", "first", "(", "raw_parser", ...
Used to get component DOI from a tag and confirm it is actually for that tag and it is not for one of its children in the list of nodenames
[ "Used", "to", "get", "component", "DOI", "from", "a", "tag", "and", "confirm", "it", "is", "actually", "for", "that", "tag", "and", "it", "is", "not", "for", "one", "of", "its", "children", "in", "the", "list", "of", "nodenames" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1486-L1507
elifesciences/elife-tools
elifetools/parseJATS.py
components
def components(soup): """ Find the components, i.e. those parts that would be assigned a unique component DOI, such as figures, tables, etc. - position is in what order the tag appears in the entire set of nodes - ordinal is in what order it is for all the tags of its own type """ components...
python
def components(soup): """ Find the components, i.e. those parts that would be assigned a unique component DOI, such as figures, tables, etc. - position is in what order the tag appears in the entire set of nodes - ordinal is in what order it is for all the tags of its own type """ components...
[ "def", "components", "(", "soup", ")", ":", "components", "=", "[", "]", "nodenames", "=", "[", "\"abstract\"", ",", "\"fig\"", ",", "\"table-wrap\"", ",", "\"media\"", ",", "\"chem-struct-wrap\"", ",", "\"sub-article\"", ",", "\"supplementary-material\"", ",", ...
Find the components, i.e. those parts that would be assigned a unique component DOI, such as figures, tables, etc. - position is in what order the tag appears in the entire set of nodes - ordinal is in what order it is for all the tags of its own type
[ "Find", "the", "components", "i", ".", "e", ".", "those", "parts", "that", "would", "be", "assigned", "a", "unique", "component", "DOI", "such", "as", "figures", "tables", "etc", ".", "-", "position", "is", "in", "what", "order", "the", "tag", "appears",...
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1509-L1695
elifesciences/elife-tools
elifetools/parseJATS.py
correspondence
def correspondence(soup): """ Find the corresp tags included in author-notes for primary correspondence """ correspondence = [] author_notes_nodes = raw_parser.author_notes(soup) if author_notes_nodes: corresp_nodes = raw_parser.corresp(author_notes_nodes) for tag in corres...
python
def correspondence(soup): """ Find the corresp tags included in author-notes for primary correspondence """ correspondence = [] author_notes_nodes = raw_parser.author_notes(soup) if author_notes_nodes: corresp_nodes = raw_parser.corresp(author_notes_nodes) for tag in corres...
[ "def", "correspondence", "(", "soup", ")", ":", "correspondence", "=", "[", "]", "author_notes_nodes", "=", "raw_parser", ".", "author_notes", "(", "soup", ")", "if", "author_notes_nodes", ":", "corresp_nodes", "=", "raw_parser", ".", "corresp", "(", "author_not...
Find the corresp tags included in author-notes for primary correspondence
[ "Find", "the", "corresp", "tags", "included", "in", "author", "-", "notes", "for", "primary", "correspondence" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1699-L1713
elifesciences/elife-tools
elifetools/parseJATS.py
author_notes
def author_notes(soup): """ Find the fn tags included in author-notes """ author_notes = [] author_notes_section = raw_parser.author_notes(soup) if author_notes_section: fn_nodes = raw_parser.fn(author_notes_section) for tag in fn_nodes: if 'fn-type' in tag.attrs: ...
python
def author_notes(soup): """ Find the fn tags included in author-notes """ author_notes = [] author_notes_section = raw_parser.author_notes(soup) if author_notes_section: fn_nodes = raw_parser.fn(author_notes_section) for tag in fn_nodes: if 'fn-type' in tag.attrs: ...
[ "def", "author_notes", "(", "soup", ")", ":", "author_notes", "=", "[", "]", "author_notes_section", "=", "raw_parser", ".", "author_notes", "(", "soup", ")", "if", "author_notes_section", ":", "fn_nodes", "=", "raw_parser", ".", "fn", "(", "author_notes_section...
Find the fn tags included in author-notes
[ "Find", "the", "fn", "tags", "included", "in", "author", "-", "notes" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1740-L1754
elifesciences/elife-tools
elifetools/parseJATS.py
full_author_notes
def full_author_notes(soup, fntype_filter=None): """ Find the fn tags included in author-notes """ notes = [] author_notes_section = raw_parser.author_notes(soup) if author_notes_section: fn_nodes = raw_parser.fn(author_notes_section) notes = footnotes(fn_nodes, fntype_filter) ...
python
def full_author_notes(soup, fntype_filter=None): """ Find the fn tags included in author-notes """ notes = [] author_notes_section = raw_parser.author_notes(soup) if author_notes_section: fn_nodes = raw_parser.fn(author_notes_section) notes = footnotes(fn_nodes, fntype_filter) ...
[ "def", "full_author_notes", "(", "soup", ",", "fntype_filter", "=", "None", ")", ":", "notes", "=", "[", "]", "author_notes_section", "=", "raw_parser", ".", "author_notes", "(", "soup", ")", "if", "author_notes_section", ":", "fn_nodes", "=", "raw_parser", "....
Find the fn tags included in author-notes
[ "Find", "the", "fn", "tags", "included", "in", "author", "-", "notes" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1757-L1768
elifesciences/elife-tools
elifetools/parseJATS.py
competing_interests
def competing_interests(soup, fntype_filter): """ Find the fn tags included in the competing interest """ competing_interests_section = extract_nodes(soup, "fn-group", attr="content-type", value="competing-interest") if not competing_interests_section: return None fn = extract_nodes(fir...
python
def competing_interests(soup, fntype_filter): """ Find the fn tags included in the competing interest """ competing_interests_section = extract_nodes(soup, "fn-group", attr="content-type", value="competing-interest") if not competing_interests_section: return None fn = extract_nodes(fir...
[ "def", "competing_interests", "(", "soup", ",", "fntype_filter", ")", ":", "competing_interests_section", "=", "extract_nodes", "(", "soup", ",", "\"fn-group\"", ",", "attr", "=", "\"content-type\"", ",", "value", "=", "\"competing-interest\"", ")", "if", "not", "...
Find the fn tags included in the competing interest
[ "Find", "the", "fn", "tags", "included", "in", "the", "competing", "interest" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1771-L1782
elifesciences/elife-tools
elifetools/parseJATS.py
author_contributions
def author_contributions(soup, fntype_filter): """ Find the fn tags included in the competing interest """ author_contributions_section = extract_nodes(soup, "fn-group", attr="content-type", value="author-contribution") if not author_contributions_section: return None fn = extract_nodes...
python
def author_contributions(soup, fntype_filter): """ Find the fn tags included in the competing interest """ author_contributions_section = extract_nodes(soup, "fn-group", attr="content-type", value="author-contribution") if not author_contributions_section: return None fn = extract_nodes...
[ "def", "author_contributions", "(", "soup", ",", "fntype_filter", ")", ":", "author_contributions_section", "=", "extract_nodes", "(", "soup", ",", "\"fn-group\"", ",", "attr", "=", "\"content-type\"", ",", "value", "=", "\"author-contribution\"", ")", "if", "not", ...
Find the fn tags included in the competing interest
[ "Find", "the", "fn", "tags", "included", "in", "the", "competing", "interest" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1806-L1817
elifesciences/elife-tools
elifetools/parseJATS.py
full_award_groups
def full_award_groups(soup): """ Find the award-group items and return a list of details """ award_groups = [] funding_group_section = extract_nodes(soup, "funding-group") # counter for auto generated id values, if required generated_id_counter = 1 for fg in funding_group_section: ...
python
def full_award_groups(soup): """ Find the award-group items and return a list of details """ award_groups = [] funding_group_section = extract_nodes(soup, "funding-group") # counter for auto generated id values, if required generated_id_counter = 1 for fg in funding_group_section: ...
[ "def", "full_award_groups", "(", "soup", ")", ":", "award_groups", "=", "[", "]", "funding_group_section", "=", "extract_nodes", "(", "soup", ",", "\"funding-group\"", ")", "# counter for auto generated id values, if required", "generated_id_counter", "=", "1", "for", "...
Find the award-group items and return a list of details
[ "Find", "the", "award", "-", "group", "items", "and", "return", "a", "list", "of", "details" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1836-L1871
elifesciences/elife-tools
elifetools/parseJATS.py
award_groups
def award_groups(soup): """ Find the award-group items and return a list of details """ award_groups = [] funding_group_section = extract_nodes(soup, "funding-group") for fg in funding_group_section: award_group_tags = extract_nodes(fg, "award-group") for ag in award_group_tag...
python
def award_groups(soup): """ Find the award-group items and return a list of details """ award_groups = [] funding_group_section = extract_nodes(soup, "funding-group") for fg in funding_group_section: award_group_tags = extract_nodes(fg, "award-group") for ag in award_group_tag...
[ "def", "award_groups", "(", "soup", ")", ":", "award_groups", "=", "[", "]", "funding_group_section", "=", "extract_nodes", "(", "soup", ",", "\"funding-group\"", ")", "for", "fg", "in", "funding_group_section", ":", "award_group_tags", "=", "extract_nodes", "(", ...
Find the award-group items and return a list of details
[ "Find", "the", "award", "-", "group", "items", "and", "return", "a", "list", "of", "details" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1875-L1896
elifesciences/elife-tools
elifetools/parseJATS.py
award_group_funding_source
def award_group_funding_source(tag): """ Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section """ award_group_funding_source = [] funding_source_tags = extract_nodes(tag, "funding-source") for t in funding_source_tag...
python
def award_group_funding_source(tag): """ Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section """ award_group_funding_source = [] funding_source_tags = extract_nodes(tag, "funding-source") for t in funding_source_tag...
[ "def", "award_group_funding_source", "(", "tag", ")", ":", "award_group_funding_source", "=", "[", "]", "funding_source_tags", "=", "extract_nodes", "(", "tag", ",", "\"funding-source\"", ")", "for", "t", "in", "funding_source_tags", ":", "award_group_funding_source", ...
Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section
[ "Given", "a", "funding", "group", "element", "Find", "the", "award", "group", "funding", "sources", "one", "for", "each", "item", "found", "in", "the", "get_funding_group", "section" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1900-L1910
elifesciences/elife-tools
elifetools/parseJATS.py
full_award_group_funding_source
def full_award_group_funding_source(tag): """ Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section """ award_group_funding_sources = [] funding_source_nodes = extract_nodes(tag, "funding-source") for funding_source_n...
python
def full_award_group_funding_source(tag): """ Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section """ award_group_funding_sources = [] funding_source_nodes = extract_nodes(tag, "funding-source") for funding_source_n...
[ "def", "full_award_group_funding_source", "(", "tag", ")", ":", "award_group_funding_sources", "=", "[", "]", "funding_source_nodes", "=", "extract_nodes", "(", "tag", ",", "\"funding-source\"", ")", "for", "funding_source_node", "in", "funding_source_nodes", ":", "awar...
Given a funding group element Find the award group funding sources, one for each item found in the get_funding_group section
[ "Given", "a", "funding", "group", "element", "Find", "the", "award", "group", "funding", "sources", "one", "for", "each", "item", "found", "in", "the", "get_funding_group", "section" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1913-L1942
elifesciences/elife-tools
elifetools/parseJATS.py
award_group_award_id
def award_group_award_id(tag): """ Find the award group award id, one for each item found in the get_funding_group section """ award_group_award_id = [] award_id_tags = extract_nodes(tag, "award-id") for t in award_id_tags: award_group_award_id.append(t.text) return award_group_a...
python
def award_group_award_id(tag): """ Find the award group award id, one for each item found in the get_funding_group section """ award_group_award_id = [] award_id_tags = extract_nodes(tag, "award-id") for t in award_id_tags: award_group_award_id.append(t.text) return award_group_a...
[ "def", "award_group_award_id", "(", "tag", ")", ":", "award_group_award_id", "=", "[", "]", "award_id_tags", "=", "extract_nodes", "(", "tag", ",", "\"award-id\"", ")", "for", "t", "in", "award_id_tags", ":", "award_group_award_id", ".", "append", "(", "t", "....
Find the award group award id, one for each item found in the get_funding_group section
[ "Find", "the", "award", "group", "award", "id", "one", "for", "each", "item", "found", "in", "the", "get_funding_group", "section" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1946-L1955
elifesciences/elife-tools
elifetools/parseJATS.py
award_group_principal_award_recipient
def award_group_principal_award_recipient(tag): """ Find the award group principal award recipient, one for each item found in the get_funding_group section """ award_group_principal_award_recipient = [] principal_award_recipients = extract_nodes(tag, "principal-award-recipient") for t in p...
python
def award_group_principal_award_recipient(tag): """ Find the award group principal award recipient, one for each item found in the get_funding_group section """ award_group_principal_award_recipient = [] principal_award_recipients = extract_nodes(tag, "principal-award-recipient") for t in p...
[ "def", "award_group_principal_award_recipient", "(", "tag", ")", ":", "award_group_principal_award_recipient", "=", "[", "]", "principal_award_recipients", "=", "extract_nodes", "(", "tag", ",", "\"principal-award-recipient\"", ")", "for", "t", "in", "principal_award_recipi...
Find the award group principal award recipient, one for each item found in the get_funding_group section
[ "Find", "the", "award", "group", "principal", "award", "recipient", "one", "for", "each", "item", "found", "in", "the", "get_funding_group", "section" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1958-L1987
elifesciences/elife-tools
elifetools/parseJATS.py
object_id_doi
def object_id_doi(tag, parent_tag_name=None): """DOI in an object-id tag found inside the tag""" doi = None object_id = None object_ids = raw_parser.object_id(tag, "doi") if object_ids: object_id = first([id_ for id_ in object_ids]) if parent_tag_name and object_id and object_id.parent.n...
python
def object_id_doi(tag, parent_tag_name=None): """DOI in an object-id tag found inside the tag""" doi = None object_id = None object_ids = raw_parser.object_id(tag, "doi") if object_ids: object_id = first([id_ for id_ in object_ids]) if parent_tag_name and object_id and object_id.parent.n...
[ "def", "object_id_doi", "(", "tag", ",", "parent_tag_name", "=", "None", ")", ":", "doi", "=", "None", "object_id", "=", "None", "object_ids", "=", "raw_parser", ".", "object_id", "(", "tag", ",", "\"doi\"", ")", "if", "object_ids", ":", "object_id", "=", ...
DOI in an object-id tag found inside the tag
[ "DOI", "in", "an", "object", "-", "id", "tag", "found", "inside", "the", "tag" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L1989-L2000
elifesciences/elife-tools
elifetools/parseJATS.py
title_tag_inspected
def title_tag_inspected(tag, parent_tag_name=None, p_parent_tag_name=None, direct_sibling_only=False): """Extract the title tag and sometimes inspect its parents""" title_tag = None if direct_sibling_only is True: for sibling_tag in tag: if sibling_tag.name and sibling_tag.name == "titl...
python
def title_tag_inspected(tag, parent_tag_name=None, p_parent_tag_name=None, direct_sibling_only=False): """Extract the title tag and sometimes inspect its parents""" title_tag = None if direct_sibling_only is True: for sibling_tag in tag: if sibling_tag.name and sibling_tag.name == "titl...
[ "def", "title_tag_inspected", "(", "tag", ",", "parent_tag_name", "=", "None", ",", "p_parent_tag_name", "=", "None", ",", "direct_sibling_only", "=", "False", ")", ":", "title_tag", "=", "None", "if", "direct_sibling_only", "is", "True", ":", "for", "sibling_ta...
Extract the title tag and sometimes inspect its parents
[ "Extract", "the", "title", "tag", "and", "sometimes", "inspect", "its", "parents" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2002-L2021
elifesciences/elife-tools
elifetools/parseJATS.py
title_text
def title_text(tag, parent_tag_name=None, p_parent_tag_name=None, direct_sibling_only=False): """Extract the text of a title tag and sometimes inspect its parents""" title = None title_tag = title_tag_inspected(tag, parent_tag_name, p_parent_tag_name, direct_sibling_only) if title_tag: title =...
python
def title_text(tag, parent_tag_name=None, p_parent_tag_name=None, direct_sibling_only=False): """Extract the text of a title tag and sometimes inspect its parents""" title = None title_tag = title_tag_inspected(tag, parent_tag_name, p_parent_tag_name, direct_sibling_only) if title_tag: title =...
[ "def", "title_text", "(", "tag", ",", "parent_tag_name", "=", "None", ",", "p_parent_tag_name", "=", "None", ",", "direct_sibling_only", "=", "False", ")", ":", "title", "=", "None", "title_tag", "=", "title_tag_inspected", "(", "tag", ",", "parent_tag_name", ...
Extract the text of a title tag and sometimes inspect its parents
[ "Extract", "the", "text", "of", "a", "title", "tag", "and", "sometimes", "inspect", "its", "parents" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2023-L2031
elifesciences/elife-tools
elifetools/parseJATS.py
boxed_text_to_image_block
def boxed_text_to_image_block(tag): "covert boxed-text to an image block containing an inline-graphic" tag_block = OrderedDict() image_content = body_block_image_content(first(raw_parser.inline_graphic(tag))) tag_block["type"] = "image" set_if_value(tag_block, "doi", doi_uri_to_doi(object_id_doi(tag...
python
def boxed_text_to_image_block(tag): "covert boxed-text to an image block containing an inline-graphic" tag_block = OrderedDict() image_content = body_block_image_content(first(raw_parser.inline_graphic(tag))) tag_block["type"] = "image" set_if_value(tag_block, "doi", doi_uri_to_doi(object_id_doi(tag...
[ "def", "boxed_text_to_image_block", "(", "tag", ")", ":", "tag_block", "=", "OrderedDict", "(", ")", "image_content", "=", "body_block_image_content", "(", "first", "(", "raw_parser", ".", "inline_graphic", "(", "tag", ")", ")", ")", "tag_block", "[", "\"type\""...
covert boxed-text to an image block containing an inline-graphic
[ "covert", "boxed", "-", "text", "to", "an", "image", "block", "containing", "an", "inline", "-", "graphic" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2077-L2092
elifesciences/elife-tools
elifetools/parseJATS.py
body_json
def body_json(soup, base_url=None): """ Get body json and then alter it with section wrapping and removing boxed-text """ body_content = body(soup, remove_key_info_box=True, base_url=base_url) # Wrap in a section if the first block is not a section if (body_content and len(body_content) > 0 and "type" i...
python
def body_json(soup, base_url=None): """ Get body json and then alter it with section wrapping and removing boxed-text """ body_content = body(soup, remove_key_info_box=True, base_url=base_url) # Wrap in a section if the first block is not a section if (body_content and len(body_content) > 0 and "type" i...
[ "def", "body_json", "(", "soup", ",", "base_url", "=", "None", ")", ":", "body_content", "=", "body", "(", "soup", ",", "remove_key_info_box", "=", "True", ",", "base_url", "=", "base_url", ")", "# Wrap in a section if the first block is not a section", "if", "(",...
Get body json and then alter it with section wrapping and removing boxed-text
[ "Get", "body", "json", "and", "then", "alter", "it", "with", "section", "wrapping", "and", "removing", "boxed", "-", "text" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2108-L2126
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_content_render
def body_block_content_render(tag, recursive=False, base_url=None): """ Render the tag as body content and call recursively if the tag has child tags """ block_content_list = [] tag_content = OrderedDict() if tag.name == "p": for block_content in body_block_paragraph_render(tag, bas...
python
def body_block_content_render(tag, recursive=False, base_url=None): """ Render the tag as body content and call recursively if the tag has child tags """ block_content_list = [] tag_content = OrderedDict() if tag.name == "p": for block_content in body_block_paragraph_render(tag, bas...
[ "def", "body_block_content_render", "(", "tag", ",", "recursive", "=", "False", ",", "base_url", "=", "None", ")", ":", "block_content_list", "=", "[", "]", "tag_content", "=", "OrderedDict", "(", ")", "if", "tag", ".", "name", "==", "\"p\"", ":", "for", ...
Render the tag as body content and call recursively if the tag has child tags
[ "Render", "the", "tag", "as", "body", "content", "and", "call", "recursively", "if", "the", "tag", "has", "child", "tags" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2175-L2232
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_paragraph_render
def body_block_paragraph_render(p_tag, html_flag=True, base_url=None): """ paragraphs may wrap some other body block content this is separated out so it can be called from more than one place """ # Configure the XML to HTML conversion preference for shorthand use below convert = lambda xml_strin...
python
def body_block_paragraph_render(p_tag, html_flag=True, base_url=None): """ paragraphs may wrap some other body block content this is separated out so it can be called from more than one place """ # Configure the XML to HTML conversion preference for shorthand use below convert = lambda xml_strin...
[ "def", "body_block_paragraph_render", "(", "p_tag", ",", "html_flag", "=", "True", ",", "base_url", "=", "None", ")", ":", "# Configure the XML to HTML conversion preference for shorthand use below", "convert", "=", "lambda", "xml_string", ":", "xml_to_html", "(", "html_f...
paragraphs may wrap some other body block content this is separated out so it can be called from more than one place
[ "paragraphs", "may", "wrap", "some", "other", "body", "block", "content", "this", "is", "separated", "out", "so", "it", "can", "be", "called", "from", "more", "than", "one", "place" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2234-L2271
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_caption_render
def body_block_caption_render(caption_tags, base_url=None): """fig and media tag captions are similar so use this common function""" caption_content = [] supplementary_material_tags = [] for block_tag in remove_doi_paragraph(caption_tags): # Note then skip p tags with supplementary-material ins...
python
def body_block_caption_render(caption_tags, base_url=None): """fig and media tag captions are similar so use this common function""" caption_content = [] supplementary_material_tags = [] for block_tag in remove_doi_paragraph(caption_tags): # Note then skip p tags with supplementary-material ins...
[ "def", "body_block_caption_render", "(", "caption_tags", ",", "base_url", "=", "None", ")", ":", "caption_content", "=", "[", "]", "supplementary_material_tags", "=", "[", "]", "for", "block_tag", "in", "remove_doi_paragraph", "(", "caption_tags", ")", ":", "# Not...
fig and media tag captions are similar so use this common function
[ "fig", "and", "media", "tag", "captions", "are", "similar", "so", "use", "this", "common", "function" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2273-L2290
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_supplementary_material_render
def body_block_supplementary_material_render(supp_tags, base_url=None): """fig and media tag caption may have supplementary material""" source_data = [] for supp_tag in supp_tags: for block_content in body_block_content_render(supp_tag, base_url=base_url): if block_content != {}: ...
python
def body_block_supplementary_material_render(supp_tags, base_url=None): """fig and media tag caption may have supplementary material""" source_data = [] for supp_tag in supp_tags: for block_content in body_block_content_render(supp_tag, base_url=base_url): if block_content != {}: ...
[ "def", "body_block_supplementary_material_render", "(", "supp_tags", ",", "base_url", "=", "None", ")", ":", "source_data", "=", "[", "]", "for", "supp_tag", "in", "supp_tags", ":", "for", "block_content", "in", "body_block_content_render", "(", "supp_tag", ",", "...
fig and media tag caption may have supplementary material
[ "fig", "and", "media", "tag", "caption", "may", "have", "supplementary", "material" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2292-L2301
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_paragraph_content
def body_block_paragraph_content(text): "for formatting of simple paragraphs of text only, and check if it is all whitespace" tag_content = OrderedDict() if text and text != '': tag_content["type"] = "paragraph" tag_content["text"] = clean_whitespace(text) return tag_content
python
def body_block_paragraph_content(text): "for formatting of simple paragraphs of text only, and check if it is all whitespace" tag_content = OrderedDict() if text and text != '': tag_content["type"] = "paragraph" tag_content["text"] = clean_whitespace(text) return tag_content
[ "def", "body_block_paragraph_content", "(", "text", ")", ":", "tag_content", "=", "OrderedDict", "(", ")", "if", "text", "and", "text", "!=", "''", ":", "tag_content", "[", "\"type\"", "]", "=", "\"paragraph\"", "tag_content", "[", "\"text\"", "]", "=", "cle...
for formatting of simple paragraphs of text only, and check if it is all whitespace
[ "for", "formatting", "of", "simple", "paragraphs", "of", "text", "only", "and", "check", "if", "it", "is", "all", "whitespace" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2303-L2309
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_image_content
def body_block_image_content(tag): "format a graphic or inline-graphic into a body block json format" image_content = OrderedDict() if tag: copy_attribute(tag.attrs, 'xlink:href', image_content, 'uri') if "uri" in image_content: # todo!! alt set_if_value(image_content...
python
def body_block_image_content(tag): "format a graphic or inline-graphic into a body block json format" image_content = OrderedDict() if tag: copy_attribute(tag.attrs, 'xlink:href', image_content, 'uri') if "uri" in image_content: # todo!! alt set_if_value(image_content...
[ "def", "body_block_image_content", "(", "tag", ")", ":", "image_content", "=", "OrderedDict", "(", ")", "if", "tag", ":", "copy_attribute", "(", "tag", ".", "attrs", ",", "'xlink:href'", ",", "image_content", ",", "'uri'", ")", "if", "\"uri\"", "in", "image_...
format a graphic or inline-graphic into a body block json format
[ "format", "a", "graphic", "or", "inline", "-", "graphic", "into", "a", "body", "block", "json", "format" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2311-L2319
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_title_label_caption
def body_block_title_label_caption(tag_content, title_value, label_value, caption_content, set_caption=True, prefer_title=False, prefer_label=False): """set the title, label and caption values in a consistent way set_caption: insert a "caption" field prefer_title: when on...
python
def body_block_title_label_caption(tag_content, title_value, label_value, caption_content, set_caption=True, prefer_title=False, prefer_label=False): """set the title, label and caption values in a consistent way set_caption: insert a "caption" field prefer_title: when on...
[ "def", "body_block_title_label_caption", "(", "tag_content", ",", "title_value", ",", "label_value", ",", "caption_content", ",", "set_caption", "=", "True", ",", "prefer_title", "=", "False", ",", "prefer_label", "=", "False", ")", ":", "set_if_value", "(", "tag_...
set the title, label and caption values in a consistent way set_caption: insert a "caption" field prefer_title: when only one value is available, set title rather than label. If False, set label rather than title
[ "set", "the", "title", "label", "and", "caption", "values", "in", "a", "consistent", "way" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2321-L2338
elifesciences/elife-tools
elifetools/parseJATS.py
body_block_attribution
def body_block_attribution(tag): "extract the attribution content for figures, tables, videos" attributions = [] if raw_parser.attrib(tag): for attrib_tag in raw_parser.attrib(tag): attributions.append(node_contents_str(attrib_tag)) if raw_parser.permissions(tag): # concatena...
python
def body_block_attribution(tag): "extract the attribution content for figures, tables, videos" attributions = [] if raw_parser.attrib(tag): for attrib_tag in raw_parser.attrib(tag): attributions.append(node_contents_str(attrib_tag)) if raw_parser.permissions(tag): # concatena...
[ "def", "body_block_attribution", "(", "tag", ")", ":", "attributions", "=", "[", "]", "if", "raw_parser", ".", "attrib", "(", "tag", ")", ":", "for", "attrib_tag", "in", "raw_parser", ".", "attrib", "(", "tag", ")", ":", "attributions", ".", "append", "(...
extract the attribution content for figures, tables, videos
[ "extract", "the", "attribution", "content", "for", "figures", "tables", "videos" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2340-L2360
elifesciences/elife-tools
elifetools/parseJATS.py
body_blocks
def body_blocks(soup): """ Note: for some reason this works and few other attempted methods work Search for certain node types, find the first nodes siblings of the same type Add the first sibling and the other siblings to a list and return them """ nodenames = body_block_nodenames() body_b...
python
def body_blocks(soup): """ Note: for some reason this works and few other attempted methods work Search for certain node types, find the first nodes siblings of the same type Add the first sibling and the other siblings to a list and return them """ nodenames = body_block_nodenames() body_b...
[ "def", "body_blocks", "(", "soup", ")", ":", "nodenames", "=", "body_block_nodenames", "(", ")", "body_block_tags", "=", "[", "]", "if", "not", "soup", ":", "return", "body_block_tags", "first_sibling_node", "=", "firstnn", "(", "soup", ".", "find_all", "(", ...
Note: for some reason this works and few other attempted methods work Search for certain node types, find the first nodes siblings of the same type Add the first sibling and the other siblings to a list and return them
[ "Note", ":", "for", "some", "reason", "this", "works", "and", "few", "other", "attempted", "methods", "work", "Search", "for", "certain", "node", "types", "find", "the", "first", "nodes", "siblings", "of", "the", "same", "type", "Add", "the", "first", "sib...
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2655-L2681
elifesciences/elife-tools
elifetools/parseJATS.py
abstract_json
def abstract_json(soup): """abstract in article json format""" abstract_tags = raw_parser.abstract(soup) abstract_json = None for tag in abstract_tags: if tag.get("abstract-type") is None: abstract_json = render_abstract_json(tag) return abstract_json
python
def abstract_json(soup): """abstract in article json format""" abstract_tags = raw_parser.abstract(soup) abstract_json = None for tag in abstract_tags: if tag.get("abstract-type") is None: abstract_json = render_abstract_json(tag) return abstract_json
[ "def", "abstract_json", "(", "soup", ")", ":", "abstract_tags", "=", "raw_parser", ".", "abstract", "(", "soup", ")", "abstract_json", "=", "None", "for", "tag", "in", "abstract_tags", ":", "if", "tag", ".", "get", "(", "\"abstract-type\"", ")", "is", "Non...
abstract in article json format
[ "abstract", "in", "article", "json", "format" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2763-L2770
elifesciences/elife-tools
elifetools/parseJATS.py
digest_json
def digest_json(soup): """digest in article json format""" abstract_tags = raw_parser.abstract(soup, abstract_type="executive-summary") abstract_json = None for tag in abstract_tags: abstract_json = render_abstract_json(tag) return abstract_json
python
def digest_json(soup): """digest in article json format""" abstract_tags = raw_parser.abstract(soup, abstract_type="executive-summary") abstract_json = None for tag in abstract_tags: abstract_json = render_abstract_json(tag) return abstract_json
[ "def", "digest_json", "(", "soup", ")", ":", "abstract_tags", "=", "raw_parser", ".", "abstract", "(", "soup", ",", "abstract_type", "=", "\"executive-summary\"", ")", "abstract_json", "=", "None", "for", "tag", "in", "abstract_tags", ":", "abstract_json", "=", ...
digest in article json format
[ "digest", "in", "article", "json", "format" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2773-L2779
elifesciences/elife-tools
elifetools/parseJATS.py
author_affiliations
def author_affiliations(author, html_flag=True): """compile author affiliations for json output""" # Configure the XML to HTML conversion preference for shorthand use below convert = lambda xml_string: xml_to_html(html_flag, xml_string) affilations = [] if author.get("affiliations"): for ...
python
def author_affiliations(author, html_flag=True): """compile author affiliations for json output""" # Configure the XML to HTML conversion preference for shorthand use below convert = lambda xml_string: xml_to_html(html_flag, xml_string) affilations = [] if author.get("affiliations"): for ...
[ "def", "author_affiliations", "(", "author", ",", "html_flag", "=", "True", ")", ":", "# Configure the XML to HTML conversion preference for shorthand use below", "convert", "=", "lambda", "xml_string", ":", "xml_to_html", "(", "html_flag", ",", "xml_string", ")", "affila...
compile author affiliations for json output
[ "compile", "author", "affiliations", "for", "json", "output" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2782-L2825
elifesciences/elife-tools
elifetools/parseJATS.py
author_json_details
def author_json_details(author, author_json, contributions, correspondence, competing_interests, equal_contributions_map, present_address_data, foot_notes_data, html_flag=True): # Configure the XML to HTML conversion preference for shorthand use below convert = la...
python
def author_json_details(author, author_json, contributions, correspondence, competing_interests, equal_contributions_map, present_address_data, foot_notes_data, html_flag=True): # Configure the XML to HTML conversion preference for shorthand use below convert = la...
[ "def", "author_json_details", "(", "author", ",", "author_json", ",", "contributions", ",", "correspondence", ",", "competing_interests", ",", "equal_contributions_map", ",", "present_address_data", ",", "foot_notes_data", ",", "html_flag", "=", "True", ")", ":", "# C...
add more author json
[ "add", "more", "author", "json" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L2963-L3002
elifesciences/elife-tools
elifetools/parseJATS.py
collab_to_group_author_key_map
def collab_to_group_author_key_map(authors): """compile a map of author collab to group-author-key""" collab_map = {} for author in authors: if author.get("collab"): collab_map[author.get("collab")] = author.get("group-author-key") return collab_map
python
def collab_to_group_author_key_map(authors): """compile a map of author collab to group-author-key""" collab_map = {} for author in authors: if author.get("collab"): collab_map[author.get("collab")] = author.get("group-author-key") return collab_map
[ "def", "collab_to_group_author_key_map", "(", "authors", ")", ":", "collab_map", "=", "{", "}", "for", "author", "in", "authors", ":", "if", "author", ".", "get", "(", "\"collab\"", ")", ":", "collab_map", "[", "author", ".", "get", "(", "\"collab\"", ")",...
compile a map of author collab to group-author-key
[ "compile", "a", "map", "of", "author", "collab", "to", "group", "-", "author", "-", "key" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3049-L3055
elifesciences/elife-tools
elifetools/parseJATS.py
map_equal_contributions
def map_equal_contributions(contributors): """assign numeric values to each unique equal-contrib id""" equal_contribution_map = {} equal_contribution_keys = [] for contributor in contributors: if contributor.get("references") and "equal-contrib" in contributor.get("references"): for ...
python
def map_equal_contributions(contributors): """assign numeric values to each unique equal-contrib id""" equal_contribution_map = {} equal_contribution_keys = [] for contributor in contributors: if contributor.get("references") and "equal-contrib" in contributor.get("references"): for ...
[ "def", "map_equal_contributions", "(", "contributors", ")", ":", "equal_contribution_map", "=", "{", "}", "equal_contribution_keys", "=", "[", "]", "for", "contributor", "in", "contributors", ":", "if", "contributor", ".", "get", "(", "\"references\"", ")", "and",...
assign numeric values to each unique equal-contrib id
[ "assign", "numeric", "values", "to", "each", "unique", "equal", "-", "contrib", "id" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3057-L3071
elifesciences/elife-tools
elifetools/parseJATS.py
authors_json
def authors_json(soup): """authors list in article json format""" authors_json_data = [] contributors_data = contributors(soup, "full") author_contributions_data = author_contributions(soup, None) author_competing_interests_data = competing_interests(soup, None) author_correspondence_data = full...
python
def authors_json(soup): """authors list in article json format""" authors_json_data = [] contributors_data = contributors(soup, "full") author_contributions_data = author_contributions(soup, None) author_competing_interests_data = competing_interests(soup, None) author_correspondence_data = full...
[ "def", "authors_json", "(", "soup", ")", ":", "authors_json_data", "=", "[", "]", "contributors_data", "=", "contributors", "(", "soup", ",", "\"full\"", ")", "author_contributions_data", "=", "author_contributions", "(", "soup", ",", "None", ")", "author_competin...
authors list in article json format
[ "authors", "list", "in", "article", "json", "format" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3088-L3141
elifesciences/elife-tools
elifetools/parseJATS.py
author_line
def author_line(soup): """take preferred names from authors json and format them into an author line""" author_line = None authors_json_data = authors_json(soup) author_names = extract_author_line_names(authors_json_data) if len(author_names) > 0: author_line = format_author_line(author_name...
python
def author_line(soup): """take preferred names from authors json and format them into an author line""" author_line = None authors_json_data = authors_json(soup) author_names = extract_author_line_names(authors_json_data) if len(author_names) > 0: author_line = format_author_line(author_name...
[ "def", "author_line", "(", "soup", ")", ":", "author_line", "=", "None", "authors_json_data", "=", "authors_json", "(", "soup", ")", "author_names", "=", "extract_author_line_names", "(", "authors_json_data", ")", "if", "len", "(", "author_names", ")", ">", "0",...
take preferred names from authors json and format them into an author line
[ "take", "preferred", "names", "from", "authors", "json", "and", "format", "them", "into", "an", "author", "line" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3143-L3150
elifesciences/elife-tools
elifetools/parseJATS.py
format_author_line
def format_author_line(author_names): """authorLine format depends on if there is 1, 2 or more than 2 authors""" author_line = None if not author_names: return author_line if len(author_names) <= 2: author_line = ", ".join(author_names) elif len(author_names) > 2: author_line...
python
def format_author_line(author_names): """authorLine format depends on if there is 1, 2 or more than 2 authors""" author_line = None if not author_names: return author_line if len(author_names) <= 2: author_line = ", ".join(author_names) elif len(author_names) > 2: author_line...
[ "def", "format_author_line", "(", "author_names", ")", ":", "author_line", "=", "None", "if", "not", "author_names", ":", "return", "author_line", "if", "len", "(", "author_names", ")", "<=", "2", ":", "author_line", "=", "\", \"", ".", "join", "(", "author_...
authorLine format depends on if there is 1, 2 or more than 2 authors
[ "authorLine", "format", "depends", "on", "if", "there", "is", "1", "2", "or", "more", "than", "2", "authors" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3165-L3174
elifesciences/elife-tools
elifetools/parseJATS.py
references_date
def references_date(year=None): "Handle year value parsing for some edge cases" date = None discriminator = None in_press = None if year and "in press" in year.lower().strip(): in_press = True elif year and re.match("^[0-9]+$", year): date = year elif year: discrimina...
python
def references_date(year=None): "Handle year value parsing for some edge cases" date = None discriminator = None in_press = None if year and "in press" in year.lower().strip(): in_press = True elif year and re.match("^[0-9]+$", year): date = year elif year: discrimina...
[ "def", "references_date", "(", "year", "=", "None", ")", ":", "date", "=", "None", "discriminator", "=", "None", "in_press", "=", "None", "if", "year", "and", "\"in press\"", "in", "year", ".", "lower", "(", ")", ".", "strip", "(", ")", ":", "in_press"...
Handle year value parsing for some edge cases
[ "Handle", "year", "value", "parsing", "for", "some", "edge", "cases" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3210-L3226
elifesciences/elife-tools
elifetools/parseJATS.py
references_json_authors
def references_json_authors(ref_authors, ref_content): "build the authors for references json here for testability" all_authors = references_authors(ref_authors) if all_authors != {}: if ref_content.get("type") in ["conference-proceeding", "journal", "other", ...
python
def references_json_authors(ref_authors, ref_content): "build the authors for references json here for testability" all_authors = references_authors(ref_authors) if all_authors != {}: if ref_content.get("type") in ["conference-proceeding", "journal", "other", ...
[ "def", "references_json_authors", "(", "ref_authors", ",", "ref_content", ")", ":", "all_authors", "=", "references_authors", "(", "ref_authors", ")", "if", "all_authors", "!=", "{", "}", ":", "if", "ref_content", ".", "get", "(", "\"type\"", ")", "in", "[", ...
build the authors for references json here for testability
[ "build", "the", "authors", "for", "references", "json", "here", "for", "testability" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3256-L3285
elifesciences/elife-tools
elifetools/parseJATS.py
convert_references_json
def convert_references_json(ref_content, soup=None): "Check for references that will not pass schema validation, fix or convert them to unknown" # Convert reference to unkonwn if still missing important values if ( (ref_content.get("type") == "other") or (ref_content.get("type") == ...
python
def convert_references_json(ref_content, soup=None): "Check for references that will not pass schema validation, fix or convert them to unknown" # Convert reference to unkonwn if still missing important values if ( (ref_content.get("type") == "other") or (ref_content.get("type") == ...
[ "def", "convert_references_json", "(", "ref_content", ",", "soup", "=", "None", ")", ":", "# Convert reference to unkonwn if still missing important values", "if", "(", "(", "ref_content", ".", "get", "(", "\"type\"", ")", "==", "\"other\"", ")", "or", "(", "ref_con...
Check for references that will not pass schema validation, fix or convert them to unknown
[ "Check", "for", "references", "that", "will", "not", "pass", "schema", "validation", "fix", "or", "convert", "them", "to", "unknown" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3480-L3507
elifesciences/elife-tools
elifetools/parseJATS.py
references_json_unknown_details
def references_json_unknown_details(ref_content, soup=None): "Extract detail value for references of type unknown" details = "" # Try adding pages values first if "pages" in ref_content: if "range" in ref_content["pages"]: details += ref_content["pages"]["range"] else: ...
python
def references_json_unknown_details(ref_content, soup=None): "Extract detail value for references of type unknown" details = "" # Try adding pages values first if "pages" in ref_content: if "range" in ref_content["pages"]: details += ref_content["pages"]["range"] else: ...
[ "def", "references_json_unknown_details", "(", "ref_content", ",", "soup", "=", "None", ")", ":", "details", "=", "\"\"", "# Try adding pages values first", "if", "\"pages\"", "in", "ref_content", ":", "if", "\"range\"", "in", "ref_content", "[", "\"pages\"", "]", ...
Extract detail value for references of type unknown
[ "Extract", "detail", "value", "for", "references", "of", "type", "unknown" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3540-L3569
elifesciences/elife-tools
elifetools/parseJATS.py
unwrap_appendix_box
def unwrap_appendix_box(json_content): """for use in removing unwanted boxed-content from appendices json""" if json_content.get("content") and len(json_content["content"]) > 0: first_block = json_content["content"][0] if (first_block.get("type") and first_block.get("type") == "box" ...
python
def unwrap_appendix_box(json_content): """for use in removing unwanted boxed-content from appendices json""" if json_content.get("content") and len(json_content["content"]) > 0: first_block = json_content["content"][0] if (first_block.get("type") and first_block.get("type") == "box" ...
[ "def", "unwrap_appendix_box", "(", "json_content", ")", ":", "if", "json_content", ".", "get", "(", "\"content\"", ")", "and", "len", "(", "json_content", "[", "\"content\"", "]", ")", ">", "0", ":", "first_block", "=", "json_content", "[", "\"content\"", "]...
for use in removing unwanted boxed-content from appendices json
[ "for", "use", "in", "removing", "unwanted", "boxed", "-", "content", "from", "appendices", "json" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/parseJATS.py#L3593-L3603
canonical-ols/acceptable
acceptable/_build_doubles.py
extract_schemas_from_file
def extract_schemas_from_file(source_path): """Extract schemas from 'source_path'. :returns: a list of ViewSchema objects on success, None if no schemas could be extracted. """ logging.info("Extracting schemas from %s", source_path) try: with open(source_path, 'r') as source_file: ...
python
def extract_schemas_from_file(source_path): """Extract schemas from 'source_path'. :returns: a list of ViewSchema objects on success, None if no schemas could be extracted. """ logging.info("Extracting schemas from %s", source_path) try: with open(source_path, 'r') as source_file: ...
[ "def", "extract_schemas_from_file", "(", "source_path", ")", ":", "logging", ".", "info", "(", "\"Extracting schemas from %s\"", ",", "source_path", ")", "try", ":", "with", "open", "(", "source_path", ",", "'r'", ")", "as", "source_file", ":", "source", "=", ...
Extract schemas from 'source_path'. :returns: a list of ViewSchema objects on success, None if no schemas could be extracted.
[ "Extract", "schemas", "from", "source_path", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_build_doubles.py#L129-L151
canonical-ols/acceptable
acceptable/_build_doubles.py
_get_simple_assignments
def _get_simple_assignments(tree): """Get simple assignments from node tree.""" result = {} for node in ast.walk(tree): if isinstance(node, ast.Assign): for target in node.targets: if isinstance(target, ast.Name): result[target.id] = node.value ret...
python
def _get_simple_assignments(tree): """Get simple assignments from node tree.""" result = {} for node in ast.walk(tree): if isinstance(node, ast.Assign): for target in node.targets: if isinstance(target, ast.Name): result[target.id] = node.value ret...
[ "def", "_get_simple_assignments", "(", "tree", ")", ":", "result", "=", "{", "}", "for", "node", "in", "ast", ".", "walk", "(", "tree", ")", ":", "if", "isinstance", "(", "node", ",", "ast", ".", "Assign", ")", ":", "for", "target", "in", "node", "...
Get simple assignments from node tree.
[ "Get", "simple", "assignments", "from", "node", "tree", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_build_doubles.py#L154-L162
canonical-ols/acceptable
acceptable/_build_doubles.py
extract_schemas_from_source
def extract_schemas_from_source(source, filename='<unknown>'): """Extract schemas from 'source'. The 'source' parameter must be a string, and should be valid python source. If 'source' is not valid python source, a SyntaxError will be raised. :returns: a list of ViewSchema objects. """ # ...
python
def extract_schemas_from_source(source, filename='<unknown>'): """Extract schemas from 'source'. The 'source' parameter must be a string, and should be valid python source. If 'source' is not valid python source, a SyntaxError will be raised. :returns: a list of ViewSchema objects. """ # ...
[ "def", "extract_schemas_from_source", "(", "source", ",", "filename", "=", "'<unknown>'", ")", ":", "# Track which acceptable services have been configured.", "acceptable_services", "=", "set", "(", ")", "# Track which acceptable views have been configured:", "acceptable_views", ...
Extract schemas from 'source'. The 'source' parameter must be a string, and should be valid python source. If 'source' is not valid python source, a SyntaxError will be raised. :returns: a list of ViewSchema objects.
[ "Extract", "schemas", "from", "source", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_build_doubles.py#L177-L289
canonical-ols/acceptable
acceptable/_build_doubles.py
render_value
def render_value(value): """Render a value, ensuring that any nested dicts are sorted by key.""" if isinstance(value, list): return '[' + ', '.join(render_value(v) for v in value) + ']' elif isinstance(value, dict): return ( '{' + ', '.join('{k!r}: {v}'.format( ...
python
def render_value(value): """Render a value, ensuring that any nested dicts are sorted by key.""" if isinstance(value, list): return '[' + ', '.join(render_value(v) for v in value) + ']' elif isinstance(value, dict): return ( '{' + ', '.join('{k!r}: {v}'.format( ...
[ "def", "render_value", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "list", ")", ":", "return", "'['", "+", "', '", ".", "join", "(", "render_value", "(", "v", ")", "for", "v", "in", "value", ")", "+", "']'", "elif", "isinstance", ...
Render a value, ensuring that any nested dicts are sorted by key.
[ "Render", "a", "value", "ensuring", "that", "any", "nested", "dicts", "are", "sorted", "by", "key", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_build_doubles.py#L292-L303
canonical-ols/acceptable
acceptable/_build_doubles.py
write_service_double_file
def write_service_double_file(target_root, service_name, rendered): """Render syntactically valid python service double code.""" target_path = os.path.join( target_root, 'snapstore_schemas', 'service_doubles', '%s.py' % service_name ) with open(target_path, 'w') as target_file: t...
python
def write_service_double_file(target_root, service_name, rendered): """Render syntactically valid python service double code.""" target_path = os.path.join( target_root, 'snapstore_schemas', 'service_doubles', '%s.py' % service_name ) with open(target_path, 'w') as target_file: t...
[ "def", "write_service_double_file", "(", "target_root", ",", "service_name", ",", "rendered", ")", ":", "target_path", "=", "os", ".", "path", ".", "join", "(", "target_root", ",", "'snapstore_schemas'", ",", "'service_doubles'", ",", "'%s.py'", "%", "service_name...
Render syntactically valid python service double code.
[ "Render", "syntactically", "valid", "python", "service", "double", "code", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_build_doubles.py#L346-L353
canonical-ols/acceptable
acceptable/util.py
clean_docstring
def clean_docstring(docstring): """Dedent docstring, special casing the first line.""" docstring = docstring.strip() if '\n' in docstring: # multiline docstring if docstring[0].isspace(): # whole docstring is indented return textwrap.dedent(docstring) else: ...
python
def clean_docstring(docstring): """Dedent docstring, special casing the first line.""" docstring = docstring.strip() if '\n' in docstring: # multiline docstring if docstring[0].isspace(): # whole docstring is indented return textwrap.dedent(docstring) else: ...
[ "def", "clean_docstring", "(", "docstring", ")", ":", "docstring", "=", "docstring", ".", "strip", "(", ")", "if", "'\\n'", "in", "docstring", ":", "# multiline docstring", "if", "docstring", "[", "0", "]", ".", "isspace", "(", ")", ":", "# whole docstring i...
Dedent docstring, special casing the first line.
[ "Dedent", "docstring", "special", "casing", "the", "first", "line", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/util.py#L24-L36
canonical-ols/acceptable
acceptable/util.py
_sort_schema
def _sort_schema(schema): """Recursively sorts a JSON schema by dict key.""" if isinstance(schema, dict): for k, v in sorted(schema.items()): if isinstance(v, dict): yield k, OrderedDict(_sort_schema(v)) elif isinstance(v, list): yield k, list(_so...
python
def _sort_schema(schema): """Recursively sorts a JSON schema by dict key.""" if isinstance(schema, dict): for k, v in sorted(schema.items()): if isinstance(v, dict): yield k, OrderedDict(_sort_schema(v)) elif isinstance(v, list): yield k, list(_so...
[ "def", "_sort_schema", "(", "schema", ")", ":", "if", "isinstance", "(", "schema", ",", "dict", ")", ":", "for", "k", ",", "v", "in", "sorted", "(", "schema", ".", "items", "(", ")", ")", ":", "if", "isinstance", "(", "v", ",", "dict", ")", ":", ...
Recursively sorts a JSON schema by dict key.
[ "Recursively", "sorts", "a", "JSON", "schema", "by", "dict", "key", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/util.py#L39-L59
canonical-ols/acceptable
acceptable/djangoutil.py
urlmap
def urlmap(patterns): """Recursively build a map of (group, name) => url patterns. Group is either the resolver namespace or app name for the url config. The urls are joined with any prefixes, and cleaned up of extraneous regex specific syntax.""" for pattern in patterns: group = getattr(p...
python
def urlmap(patterns): """Recursively build a map of (group, name) => url patterns. Group is either the resolver namespace or app name for the url config. The urls are joined with any prefixes, and cleaned up of extraneous regex specific syntax.""" for pattern in patterns: group = getattr(p...
[ "def", "urlmap", "(", "patterns", ")", ":", "for", "pattern", "in", "patterns", ":", "group", "=", "getattr", "(", "pattern", ",", "'namespace'", ",", "None", ")", "if", "group", "is", "None", ":", "group", "=", "getattr", "(", "pattern", ",", "'app_na...
Recursively build a map of (group, name) => url patterns. Group is either the resolver namespace or app name for the url config. The urls are joined with any prefixes, and cleaned up of extraneous regex specific syntax.
[ "Recursively", "build", "a", "map", "of", "(", "group", "name", ")", "=", ">", "url", "patterns", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/djangoutil.py#L64-L81
canonical-ols/acceptable
acceptable/djangoutil.py
get_field_schema
def get_field_schema(name, field): """Returns a JSON Schema representation of a form field.""" field_schema = { 'type': 'string', } if field.label: field_schema['title'] = str(field.label) # force translation if field.help_text: field_schema['description'] = str(field.help...
python
def get_field_schema(name, field): """Returns a JSON Schema representation of a form field.""" field_schema = { 'type': 'string', } if field.label: field_schema['title'] = str(field.label) # force translation if field.help_text: field_schema['description'] = str(field.help...
[ "def", "get_field_schema", "(", "name", ",", "field", ")", ":", "field_schema", "=", "{", "'type'", ":", "'string'", ",", "}", "if", "field", ".", "label", ":", "field_schema", "[", "'title'", "]", "=", "str", "(", "field", ".", "label", ")", "# force ...
Returns a JSON Schema representation of a form field.
[ "Returns", "a", "JSON", "Schema", "representation", "of", "a", "form", "field", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/djangoutil.py#L84-L127
canonical-ols/acceptable
acceptable/djangoutil.py
get_form_schema
def get_form_schema(form): """Return a JSON Schema object for a Django Form.""" schema = { 'type': 'object', 'properties': {}, } for name, field in form.base_fields.items(): schema['properties'][name] = get_field_schema(name, field) if field.required: schema....
python
def get_form_schema(form): """Return a JSON Schema object for a Django Form.""" schema = { 'type': 'object', 'properties': {}, } for name, field in form.base_fields.items(): schema['properties'][name] = get_field_schema(name, field) if field.required: schema....
[ "def", "get_form_schema", "(", "form", ")", ":", "schema", "=", "{", "'type'", ":", "'object'", ",", "'properties'", ":", "{", "}", ",", "}", "for", "name", ",", "field", "in", "form", ".", "base_fields", ".", "items", "(", ")", ":", "schema", "[", ...
Return a JSON Schema object for a Django Form.
[ "Return", "a", "JSON", "Schema", "object", "for", "a", "Django", "Form", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/djangoutil.py#L130-L142
canonical-ols/acceptable
acceptable/djangoutil.py
DjangoAPI.handler
def handler(self, handler_class): """Link to an API handler class (e.g. piston or DRF).""" self.handler_class = handler_class # we take the docstring from the handler class, not the methods if self.docs is None and handler_class.__doc__: self.docs = clean_docstring(handler_cl...
python
def handler(self, handler_class): """Link to an API handler class (e.g. piston or DRF).""" self.handler_class = handler_class # we take the docstring from the handler class, not the methods if self.docs is None and handler_class.__doc__: self.docs = clean_docstring(handler_cl...
[ "def", "handler", "(", "self", ",", "handler_class", ")", ":", "self", ".", "handler_class", "=", "handler_class", "# we take the docstring from the handler class, not the methods", "if", "self", ".", "docs", "is", "None", "and", "handler_class", ".", "__doc__", ":", ...
Link to an API handler class (e.g. piston or DRF).
[ "Link", "to", "an", "API", "handler", "class", "(", "e", ".", "g", ".", "piston", "or", "DRF", ")", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/djangoutil.py#L217-L223
elifesciences/elife-tools
elifetools/utils_html.py
xml_to_html
def xml_to_html(html_flag, xml_string, base_url=None): "For formatting json output into HTML friendly format" if not xml_string or not html_flag is True: return xml_string html_string = xml_string html_string = remove_comment_tags(html_string) # Escape unmatched angle brackets if '<' in...
python
def xml_to_html(html_flag, xml_string, base_url=None): "For formatting json output into HTML friendly format" if not xml_string or not html_flag is True: return xml_string html_string = xml_string html_string = remove_comment_tags(html_string) # Escape unmatched angle brackets if '<' in...
[ "def", "xml_to_html", "(", "html_flag", ",", "xml_string", ",", "base_url", "=", "None", ")", ":", "if", "not", "xml_string", "or", "not", "html_flag", "is", "True", ":", "return", "xml_string", "html_string", "=", "xml_string", "html_string", "=", "remove_com...
For formatting json output into HTML friendly format
[ "For", "formatting", "json", "output", "into", "HTML", "friendly", "format" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/utils_html.py#L5-L29
elifesciences/elife-tools
elifetools/utils_html.py
replace_simple_tags
def replace_simple_tags(s, from_tag='italic', to_tag='i', to_open_tag=None): """ Replace tags such as <italic> to <i> This does not validate markup """ if to_open_tag: s = s.replace('<' + from_tag + '>', to_open_tag) elif to_tag: s = s.replace('<' + from_tag + '>', '<' + to_tag +...
python
def replace_simple_tags(s, from_tag='italic', to_tag='i', to_open_tag=None): """ Replace tags such as <italic> to <i> This does not validate markup """ if to_open_tag: s = s.replace('<' + from_tag + '>', to_open_tag) elif to_tag: s = s.replace('<' + from_tag + '>', '<' + to_tag +...
[ "def", "replace_simple_tags", "(", "s", ",", "from_tag", "=", "'italic'", ",", "to_tag", "=", "'i'", ",", "to_open_tag", "=", "None", ")", ":", "if", "to_open_tag", ":", "s", "=", "s", ".", "replace", "(", "'<'", "+", "from_tag", "+", "'>'", ",", "to...
Replace tags such as <italic> to <i> This does not validate markup
[ "Replace", "tags", "such", "as", "<italic", ">", "to", "<i", ">", "This", "does", "not", "validate", "markup" ]
train
https://github.com/elifesciences/elife-tools/blob/4b9e38cbe485c61a4ed7cbd8970c6b318334fd86/elifetools/utils_html.py#L70-L89
canonical-ols/acceptable
acceptable/_validation.py
validate_body
def validate_body(schema): """Validate the body of incoming requests for a flask view. An example usage might look like this:: from snapstore_schemas import validate_body @validate_body({ 'type': 'array', 'items': { 'type': 'object', 'p...
python
def validate_body(schema): """Validate the body of incoming requests for a flask view. An example usage might look like this:: from snapstore_schemas import validate_body @validate_body({ 'type': 'array', 'items': { 'type': 'object', 'p...
[ "def", "validate_body", "(", "schema", ")", ":", "location", "=", "get_callsite_location", "(", ")", "def", "decorator", "(", "fn", ")", ":", "validate_schema", "(", "schema", ")", "wrapper", "=", "wrap_request", "(", "fn", ",", "schema", ")", "record_schema...
Validate the body of incoming requests for a flask view. An example usage might look like this:: from snapstore_schemas import validate_body @validate_body({ 'type': 'array', 'items': { 'type': 'object', 'properties': { ...
[ "Validate", "the", "body", "of", "incoming", "requests", "for", "a", "flask", "view", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_validation.py#L30-L81
canonical-ols/acceptable
acceptable/_validation.py
record_schemas
def record_schemas( fn, wrapper, location, request_schema=None, response_schema=None): """Support extracting the schema from the decorated function.""" # have we already been decorated by an acceptable api call? has_acceptable = hasattr(fn, '_acceptable_metadata') if request_schema is not None:...
python
def record_schemas( fn, wrapper, location, request_schema=None, response_schema=None): """Support extracting the schema from the decorated function.""" # have we already been decorated by an acceptable api call? has_acceptable = hasattr(fn, '_acceptable_metadata') if request_schema is not None:...
[ "def", "record_schemas", "(", "fn", ",", "wrapper", ",", "location", ",", "request_schema", "=", "None", ",", "response_schema", "=", "None", ")", ":", "# have we already been decorated by an acceptable api call?", "has_acceptable", "=", "hasattr", "(", "fn", ",", "...
Support extracting the schema from the decorated function.
[ "Support", "extracting", "the", "schema", "from", "the", "decorated", "function", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_validation.py#L108-L128
canonical-ols/acceptable
acceptable/_validation.py
validate_output
def validate_output(schema): """Validate the body of a response from a flask view. Like `validate_body`, this function compares a json document to a jsonschema specification. However, this function applies the schema to the view response. Instead of the view returning a flask response object, it s...
python
def validate_output(schema): """Validate the body of a response from a flask view. Like `validate_body`, this function compares a json document to a jsonschema specification. However, this function applies the schema to the view response. Instead of the view returning a flask response object, it s...
[ "def", "validate_output", "(", "schema", ")", ":", "location", "=", "get_callsite_location", "(", ")", "def", "decorator", "(", "fn", ")", ":", "validate_schema", "(", "schema", ")", "wrapper", "=", "wrap_response", "(", "fn", ",", "schema", ")", "record_sch...
Validate the body of a response from a flask view. Like `validate_body`, this function compares a json document to a jsonschema specification. However, this function applies the schema to the view response. Instead of the view returning a flask response object, it should instead return a Python li...
[ "Validate", "the", "body", "of", "a", "response", "from", "a", "flask", "view", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_validation.py#L131-L167
canonical-ols/acceptable
acceptable/_validation.py
validate
def validate(payload, schema): """Validate `payload` against `schema`, returning an error list. jsonschema provides lots of information in it's errors, but it can be a bit of work to extract all the information. """ v = jsonschema.Draft4Validator( schema, format_checker=jsonschema.FormatChe...
python
def validate(payload, schema): """Validate `payload` against `schema`, returning an error list. jsonschema provides lots of information in it's errors, but it can be a bit of work to extract all the information. """ v = jsonschema.Draft4Validator( schema, format_checker=jsonschema.FormatChe...
[ "def", "validate", "(", "payload", ",", "schema", ")", ":", "v", "=", "jsonschema", ".", "Draft4Validator", "(", "schema", ",", "format_checker", "=", "jsonschema", ".", "FormatChecker", "(", ")", ")", "error_list", "=", "[", "]", "for", "error", "in", "...
Validate `payload` against `schema`, returning an error list. jsonschema provides lots of information in it's errors, but it can be a bit of work to extract all the information.
[ "Validate", "payload", "against", "schema", "returning", "an", "error", "list", "." ]
train
https://github.com/canonical-ols/acceptable/blob/6ccbe969078166a5315d857da38b59b43b29fadc/acceptable/_validation.py#L207-L220
lalinsky/python-phoenixdb
phoenixdb/__init__.py
connect
def connect(url, max_retries=None, **kwargs): """Connects to a Phoenix query server. :param url: URL to the Phoenix query server, e.g. ``http://localhost:8765/`` :param autocommit: Switch the connection to autocommit mode. :param readonly: Switch the connection to readonly mod...
python
def connect(url, max_retries=None, **kwargs): """Connects to a Phoenix query server. :param url: URL to the Phoenix query server, e.g. ``http://localhost:8765/`` :param autocommit: Switch the connection to autocommit mode. :param readonly: Switch the connection to readonly mod...
[ "def", "connect", "(", "url", ",", "max_retries", "=", "None", ",", "*", "*", "kwargs", ")", ":", "client", "=", "AvaticaClient", "(", "url", ",", "max_retries", "=", "max_retries", ")", "client", ".", "connect", "(", ")", "return", "Connection", "(", ...
Connects to a Phoenix query server. :param url: URL to the Phoenix query server, e.g. ``http://localhost:8765/`` :param autocommit: Switch the connection to autocommit mode. :param readonly: Switch the connection to readonly mode. :param max_retries: The maximum numbe...
[ "Connects", "to", "a", "Phoenix", "query", "server", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/__init__.py#L44-L67
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.connect
def connect(self): """Opens a HTTP connection to the RPC server.""" logger.debug("Opening connection to %s:%s", self.url.hostname, self.url.port) try: self.connection = httplib.HTTPConnection(self.url.hostname, self.url.port) self.connection.connect() except (http...
python
def connect(self): """Opens a HTTP connection to the RPC server.""" logger.debug("Opening connection to %s:%s", self.url.hostname, self.url.port) try: self.connection = httplib.HTTPConnection(self.url.hostname, self.url.port) self.connection.connect() except (http...
[ "def", "connect", "(", "self", ")", ":", "logger", ".", "debug", "(", "\"Opening connection to %s:%s\"", ",", "self", ".", "url", ".", "hostname", ",", "self", ".", "url", ".", "port", ")", "try", ":", "self", ".", "connection", "=", "httplib", ".", "H...
Opens a HTTP connection to the RPC server.
[ "Opens", "a", "HTTP", "connection", "to", "the", "RPC", "server", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L151-L158
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.close
def close(self): """Closes the HTTP connection to the RPC server.""" if self.connection is not None: logger.debug("Closing connection to %s:%s", self.url.hostname, self.url.port) try: self.connection.close() except httplib.HTTPException: ...
python
def close(self): """Closes the HTTP connection to the RPC server.""" if self.connection is not None: logger.debug("Closing connection to %s:%s", self.url.hostname, self.url.port) try: self.connection.close() except httplib.HTTPException: ...
[ "def", "close", "(", "self", ")", ":", "if", "self", ".", "connection", "is", "not", "None", ":", "logger", ".", "debug", "(", "\"Closing connection to %s:%s\"", ",", "self", ".", "url", ".", "hostname", ",", "self", ".", "url", ".", "port", ")", "try"...
Closes the HTTP connection to the RPC server.
[ "Closes", "the", "HTTP", "connection", "to", "the", "RPC", "server", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L160-L168
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.connection_sync
def connection_sync(self, connection_id, connProps=None): """Synchronizes connection properties with the server. :param connection_id: ID of the current connection. :param connProps: Dictionary with the properties that should be changed. :returns: A...
python
def connection_sync(self, connection_id, connProps=None): """Synchronizes connection properties with the server. :param connection_id: ID of the current connection. :param connProps: Dictionary with the properties that should be changed. :returns: A...
[ "def", "connection_sync", "(", "self", ",", "connection_id", ",", "connProps", "=", "None", ")", ":", "if", "connProps", "is", "None", ":", "connProps", "=", "{", "}", "request", "=", "requests_pb2", ".", "ConnectionSyncRequest", "(", ")", "request", ".", ...
Synchronizes connection properties with the server. :param connection_id: ID of the current connection. :param connProps: Dictionary with the properties that should be changed. :returns: A ``common_pb2.ConnectionProperties`` object.
[ "Synchronizes", "connection", "properties", "with", "the", "server", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L286-L314
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.open_connection
def open_connection(self, connection_id, info=None): """Opens a new connection. :param connection_id: ID of the connection to open. """ request = requests_pb2.OpenConnectionRequest() request.connection_id = connection_id if info is not None: # Inf...
python
def open_connection(self, connection_id, info=None): """Opens a new connection. :param connection_id: ID of the connection to open. """ request = requests_pb2.OpenConnectionRequest() request.connection_id = connection_id if info is not None: # Inf...
[ "def", "open_connection", "(", "self", ",", "connection_id", ",", "info", "=", "None", ")", ":", "request", "=", "requests_pb2", ".", "OpenConnectionRequest", "(", ")", "request", ".", "connection_id", "=", "connection_id", "if", "info", "is", "not", "None", ...
Opens a new connection. :param connection_id: ID of the connection to open.
[ "Opens", "a", "new", "connection", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L316-L331
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.close_connection
def close_connection(self, connection_id): """Closes a connection. :param connection_id: ID of the connection to close. """ request = requests_pb2.CloseConnectionRequest() request.connection_id = connection_id self._apply(request)
python
def close_connection(self, connection_id): """Closes a connection. :param connection_id: ID of the connection to close. """ request = requests_pb2.CloseConnectionRequest() request.connection_id = connection_id self._apply(request)
[ "def", "close_connection", "(", "self", ",", "connection_id", ")", ":", "request", "=", "requests_pb2", ".", "CloseConnectionRequest", "(", ")", "request", ".", "connection_id", "=", "connection_id", "self", ".", "_apply", "(", "request", ")" ]
Closes a connection. :param connection_id: ID of the connection to close.
[ "Closes", "a", "connection", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L333-L341
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.create_statement
def create_statement(self, connection_id): """Creates a new statement. :param connection_id: ID of the current connection. :returns: New statement ID. """ request = requests_pb2.CreateStatementRequest() request.connection_id = connection_id ...
python
def create_statement(self, connection_id): """Creates a new statement. :param connection_id: ID of the current connection. :returns: New statement ID. """ request = requests_pb2.CreateStatementRequest() request.connection_id = connection_id ...
[ "def", "create_statement", "(", "self", ",", "connection_id", ")", ":", "request", "=", "requests_pb2", ".", "CreateStatementRequest", "(", ")", "request", ".", "connection_id", "=", "connection_id", "response_data", "=", "self", ".", "_apply", "(", "request", "...
Creates a new statement. :param connection_id: ID of the current connection. :returns: New statement ID.
[ "Creates", "a", "new", "statement", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L343-L358
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.close_statement
def close_statement(self, connection_id, statement_id): """Closes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to close. """ request = requests_pb2.CloseStatementRequest() request.conne...
python
def close_statement(self, connection_id, statement_id): """Closes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to close. """ request = requests_pb2.CloseStatementRequest() request.conne...
[ "def", "close_statement", "(", "self", ",", "connection_id", ",", "statement_id", ")", ":", "request", "=", "requests_pb2", ".", "CloseStatementRequest", "(", ")", "request", ".", "connection_id", "=", "connection_id", "request", ".", "statement_id", "=", "stateme...
Closes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to close.
[ "Closes", "a", "statement", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L360-L373
lalinsky/python-phoenixdb
phoenixdb/avatica/client.py
AvaticaClient.prepare_and_execute
def prepare_and_execute(self, connection_id, statement_id, sql, max_rows_total=None, first_frame_max_size=None): """Prepares and immediately executes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to prepare. ...
python
def prepare_and_execute(self, connection_id, statement_id, sql, max_rows_total=None, first_frame_max_size=None): """Prepares and immediately executes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to prepare. ...
[ "def", "prepare_and_execute", "(", "self", ",", "connection_id", ",", "statement_id", ",", "sql", ",", "max_rows_total", "=", "None", ",", "first_frame_max_size", "=", "None", ")", ":", "request", "=", "requests_pb2", ".", "PrepareAndExecuteRequest", "(", ")", "...
Prepares and immediately executes a statement. :param connection_id: ID of the current connection. :param statement_id: ID of the statement to prepare. :param sql: SQL query. :param max_rows_total: The maximum number of rows that will b...
[ "Prepares", "and", "immediately", "executes", "a", "statement", "." ]
train
https://github.com/lalinsky/python-phoenixdb/blob/1bb34488dd530ca65f91b29ef16aa7b71f26b806/phoenixdb/avatica/client.py#L375-L408