Szymon Woźniak commited on
Commit
7140424
1 Parent(s): 6bcdbf9

add descriptions to language typology and cross-domain pages

Browse files
pages/3_Language_Typology.py CHANGED
@@ -14,6 +14,11 @@ def get_typology_df():
14
  st.set_page_config(page_title="Language Typology", page_icon="📈")
15
 
16
  st.markdown("# Language Typology")
 
 
 
 
 
17
 
18
  df = get_typology_df()
19
 
 
14
  st.set_page_config(page_title="Language Typology", page_icon="📈")
15
 
16
  st.markdown("# Language Typology")
17
+ st.write("""\
18
+ Languages can be described using hundreds of linguistic features. [World Atlas of Language Structures](https://doi.org/10.5281/zenodo.7385533) lists almost 200 different features. Since our work focuses on sentiment classification, we select 10 features that seem to be the most relevant to the task of sentiment expression.
19
+
20
+ The table below presents the languages included in the MMS corpus, their family, genus and their values for the selected linguistic features.
21
+ You can use the **'Add filters'** checkbox to filter the table by any of the columns.""")
22
 
23
  df = get_typology_df()
24
 
pages/{4_Domain_Transfer.py → 4_Cross-domain_Transfer.py} RENAMED
@@ -17,6 +17,10 @@ st.markdown(f"# {TITLE}")
17
 
18
  df = get_domain_transfer_data()
19
 
 
 
 
 
20
  st.plotly_chart(
21
  px.box(
22
  df,
 
17
 
18
  df = get_domain_transfer_data()
19
 
20
+ st.write("""\
21
+ The figure below shows the results of an experiment comparing the effectiveness of transfer learning between data modalities. As expected, when models are tested against datasets from the same domain (news, social media, reviews), the average performance is much higher than on out-of-domain datasets. What is striking is the visible deterioration of the performance of models trained in the news domain. When knowledge transfer happens between social media and review domains, the average performance of models stays relatively the same."""
22
+ )
23
+
24
  st.plotly_chart(
25
  px.box(
26
  df,