File size: 83,069 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
{
    "paper_id": "2021",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T01:13:10.745125Z"
    },
    "title": "Leveraging English Word Embeddings for Semi-Automatic Semantic Classification in N\u00eahiyaw\u00eawin (Plains Cree)",
    "authors": [
        {
            "first": "Atticus",
            "middle": [
                "G"
            ],
            "last": "Harrigan",
            "suffix": "",
            "affiliation": {},
            "email": "atticus.harrigan@ualberta.ca"
        },
        {
            "first": "Antti",
            "middle": [],
            "last": "Arppe",
            "suffix": "",
            "affiliation": {},
            "email": "arppe@ualberta.ca"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "This paper details a semi-automatic method of word clustering for the Algonquian language, N\u00eahiyaw\u00eawin (Plains Cree). Although this method worked well, particularly for nouns, it required some amount of manual postprocessing. The main benefit of this approach over implementing an existing classification ontology is that this method approaches the language from an endogenous point of view, while performing classification quicker than in a fully manual context. 1 There is one attempt at semantically classifying N\u00eahiyaw\u00eawin through automatic means found in Dacanay et al. (2021). This work makes use of similar techniques as desccribed in this paper, differing mainly in its mapping of N\u00eahiyaw\u00eawin words onto Wordnet classes.",
    "pdf_parse": {
        "paper_id": "2021",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "This paper details a semi-automatic method of word clustering for the Algonquian language, N\u00eahiyaw\u00eawin (Plains Cree). Although this method worked well, particularly for nouns, it required some amount of manual postprocessing. The main benefit of this approach over implementing an existing classification ontology is that this method approaches the language from an endogenous point of view, while performing classification quicker than in a fully manual context. 1 There is one attempt at semantically classifying N\u00eahiyaw\u00eawin through automatic means found in Dacanay et al. (2021). This work makes use of similar techniques as desccribed in this paper, differing mainly in its mapping of N\u00eahiyaw\u00eawin words onto Wordnet classes.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Grouping words into semantic subclasses within a part of speech is a technique used widely throughout quantitative and predictive studies in the field of linguistics. Bresnan et al. (2007) use high level verb classes to predict the English dative alternation, Arppe et al. (2008) uses verb class as one of the feature sets to help predict the alternation of Finnish think verbs, and Yu et al. (2017) use polarity classifications (good vs bad) from pre-defined lexica such as WordNet (Miller, 1998) . In many cases, classifications within word classes allow researchers to group words into smaller cohesive groups to allow for use as predictors in modelling. Rather than using thousands individual lexemes as predictors, one can use a word's class to generalize over the semantic features of individual lexemes to allow for significantly more statistical power.",
                "cite_spans": [
                    {
                        "start": 167,
                        "end": 188,
                        "text": "Bresnan et al. (2007)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 260,
                        "end": 279,
                        "text": "Arppe et al. (2008)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 383,
                        "end": 399,
                        "text": "Yu et al. (2017)",
                        "ref_id": "BIBREF22"
                    },
                    {
                        "start": 483,
                        "end": 497,
                        "text": "(Miller, 1998)",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "While extensive ontologies of word classifications exist for majority languages like English (Miller, 1998) , German (Hamp and Feldweg, 1997) , and Chinese (Wang and Bond, 2013) , minority languages, especially lesser resourced languages in North America generally do not boast such resources. 1 Where such ontologies do exist, for ex-ample in Innu-aimun (Eastern Cree) (Visitor et al., 2013) , they are often manually created, an expensive process in terms of time. Alternatively, they may be based upon English ontologies such as WordNet. This opens the window to near-automatic ontology creation by associating definitions in a target language and English through a variety of methods. This is especially important, given the amount of time and effort that goes into manually classifying a lexicon through either an existing ontology (be it something like Rapidwords 2 or even Levin's like classes (Levin, 1993) ). Moreover, there is a motivation based in understanding a language and its lexicalization process on its own terms, though how to do this with a lesser resourced language remains unclear.",
                "cite_spans": [
                    {
                        "start": 93,
                        "end": 107,
                        "text": "(Miller, 1998)",
                        "ref_id": "BIBREF15"
                    },
                    {
                        "start": 117,
                        "end": 141,
                        "text": "(Hamp and Feldweg, 1997)",
                        "ref_id": "BIBREF7"
                    },
                    {
                        "start": 156,
                        "end": 177,
                        "text": "(Wang and Bond, 2013)",
                        "ref_id": "BIBREF19"
                    },
                    {
                        "start": 294,
                        "end": 295,
                        "text": "1",
                        "ref_id": null
                    },
                    {
                        "start": 370,
                        "end": 392,
                        "text": "(Visitor et al., 2013)",
                        "ref_id": "BIBREF18"
                    },
                    {
                        "start": 901,
                        "end": 914,
                        "text": "(Levin, 1993)",
                        "ref_id": "BIBREF9"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "We begun word classification in preparation for modelling a morpho-syntactic alternation in N\u00eahiyaw\u00eawin verbs. One hypothesis we developed for this alternation, based on Arppe et al. (2008) , is that the semantic classes of the verbs themselves as well as their nominal arguments would inform the verbal alternation. Due to constraints of time, we investigated methods to automatically classify both verbs and nouns in N\u00eahiyaw\u00eawin. Although statistical modelling remains the immediate motivator for the authors, semantic/thematic classifications have a wide range of benefits for language learners and revitalization, particularly in online lexicographic resources, where one may want to view all words to do with a theme, rather than simply finding translations of single English words.",
                "cite_spans": [
                    {
                        "start": 170,
                        "end": 189,
                        "text": "Arppe et al. (2008)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Background",
                "sec_num": "2"
            },
            {
                "text": "In creating a framework for automatic semantic classification we make use of Word2vec (Mikolov et al., 2013a) word embeddings. Word embeddings are words represented by n-dimensional vectors. These vectors are ultimately derived from a word's context in some corpus through the Word2vec algorithm. Unfortunately, the Word2vec method is sensitive to corpus size. We initially attempted to create basic word and feature co-occurrence matrices based on a 140,000 token N\u00eahiyaw\u00eawin corpus (Arppe et al., 2020) to create word vectors using Principal Components Analysis, but in the end found the results to be not practically useful. Similarly, an attempt at both tf-idf and Word2Vec using only the N\u00eahiyaw\u00eawin dictionary produces mostly ill-formed groupings, though in these cases preprocessing by splitting verbs and nouns was not performed. Regardless, the poor performance was most certainly due simply to the paucity of data. Although the available corpora are small, N\u00eahiyaw\u00eawin does have several English-to-N\u00eahiyaw\u00eawin dictionaries, the largest being Wolvengrey (2001) . Although a bilingual N\u00eahiyaw\u00eawin-English dictionary, it is one formed from an Indigenous point of view, based on vocabulary from previous dictionaries, some of which have been compiled by N\u00eahiyaw\u00eawin communities from their own perspectives, or gleaned from a number of texts collections rather than attempting to find N\u00eahiyaw\u00eawin word matches for a pre-defined set of English words. This results in dictionary entries such as sakapw\u00eaw: it roasts over a fire (by hanging, with string on stick). Definitions such as this take into account the nuanced cultural understanding reflected in the word's morphology.",
                "cite_spans": [
                    {
                        "start": 86,
                        "end": 109,
                        "text": "(Mikolov et al., 2013a)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 484,
                        "end": 504,
                        "text": "(Arppe et al., 2020)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 1052,
                        "end": 1069,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Background",
                "sec_num": "2"
            },
            {
                "text": "To address the issue of corpus size, we attempted to bootstrap our classification scheme with pretrained English vectors in the form of the 3 million word Google News Corpus, which represents every word with a 300-dimensional vector. 3 We make use of the English definitions (sometimes also referred to as glosses) provided in Wolvengrey (2001) and fit to each word its respective Google News Corpus vector. This dictionary makes use of lemmas as headwords, and contains 21,717 entries. The presumption is that the real-world referents (at least in terms of denotation) of English and N\u00eahiyaw\u00eawin words are approximately comparable, in particular when taking the entire set of words in an English definition. Stop words were removed, and where content words were present in definitions in Wolvengrey (2001) but not available in the Google News Corpus, synonyms were used (one such example might be the word mit\u00eawin, which is unavailable in the corpus and thus would replaced with something like medicine lodge or deleted if a synonym was given in the definition as well). Because the Google News Corpus is based in American spelling, while Wolvengrey (2001) is based in Canadian spelling, American forms (e.g. color, gray) were converted into Canadian forms (e.g. colour, grey). If such preprocessing is not performed, these words are simply unavailable for clustering, as they lack a matching vector. 4 Where a N\u00eahiyaw\u00eawin word had more than one word sense, each sense was given a separate entry and the second entry was marked with a unique identifier. Finally, where needed, words in the N\u00eahiyaw\u00eawin definitions were lemmatized.",
                "cite_spans": [
                    {
                        "start": 234,
                        "end": 235,
                        "text": "3",
                        "ref_id": null
                    },
                    {
                        "start": 327,
                        "end": 344,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 789,
                        "end": 806,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 1140,
                        "end": 1157,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 1402,
                        "end": 1403,
                        "text": "4",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "Once every word in Wolvengrey (2001) definitions matched an entry in the Google News Corpus, we associated each word in a N\u00eahiyaw\u00eawin definition with its respective Google News Vector. That is, given a definition such as aw\u00e2sisihk\u00e2nis: small doll, the resulting structure would be:",
                "cite_spans": [
                    {
                        "start": 19,
                        "end": 36,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "aw\u00e2sisihk\u00e2nis = \uf8ee \uf8ef \uf8ef \uf8ef \uf8f0 0.159 0.096 \u22120.125 . . . \uf8f9 \uf8fa \uf8fa \uf8fa \uf8fb \uf8ee \uf8ef \uf8ef \uf8ef \uf8f0 0.108 0.031 \u22120.034 . . . \uf8f9 \uf8fa \uf8fa \uf8fa \uf8fb",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "Because all word-vectors in the Google News Corpus are of the same dimensionality, we then took the resulting definition and averaged, per dimension, the values of all its constituent word-vectors. This produced a single 300-dimensional vector that acts as a sort of naive sentence vector for each of the English glosses/definitions:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "aw\u00e2sisihk\u00e2nis = \uf8ee \uf8ef \uf8ef \uf8ef \uf8f0 0.134 0.064 \u22120.080 . . . \uf8f9 \uf8fa \uf8fa \uf8fa \uf8fb",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "Mikolov et al. (2013b) mention this sort of naive representation and suggests the use of phrase vectors instead of word vectors to address the representation of non-compositional idioms; however, given the way Wolvengrey (2001) 's definitions are written (e.g. with few idiomatic or metaphorical constructions), and for reasons of computational simplicity, we opted to use the above naive implementation in this paper. After creating the sentence (or English definition) vectors, we proceeded to cluster definitions with similar vectors together. To achieve this, we created a Euclidean distance matrix from the sentence vectors and made use of the hclust package in R (R Core Team, 2017) to preform hierarchical agglomerative clustering using the Ward method (based on the experience of (Arppe et al., 2008) in using the method to produce multiple levels of smaller, spherical clusters). This form of clustering is essentially a bottom-up approach where groupings are made by starting with individual labels with the shortest distance, then iteratively at a higher level making use of the clusters that result from the previous step or remaining individual levels; this second step is repeated until there is a single cluster containing all labels. This method of clustering creates a cluster tree that can be cut at any specified level after the analysis has been completed to select different numbers of clusters, allowing researchers some degree of flexibility without needing to rerun the clustering. This method is very similar to what has been done by both Arppe et al. (2008) , Bresnan et al. (2007) , and Divjak and Gries (2006) . The choice of what number of clusters was made based on an evaluation of the effectiveness of the clusters, based on an impressionistic overview by the authors.",
                "cite_spans": [
                    {
                        "start": 210,
                        "end": 227,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 788,
                        "end": 808,
                        "text": "(Arppe et al., 2008)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 1564,
                        "end": 1583,
                        "text": "Arppe et al. (2008)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 1586,
                        "end": 1607,
                        "text": "Bresnan et al. (2007)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 1614,
                        "end": 1637,
                        "text": "Divjak and Gries (2006)",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "For our purposes, we focused on the semantic classification of N\u00eahiyaw\u00eawin nouns and verbs. N\u00eahiyaw\u00eawin verbs are naturally morphosemantically divided into four separate classes: Intransitive verbs with a single inanimate argument (VII), Intransitive verbs with a single animate argument (VAI), transitive verbs with an animate actor 5 and an inanimate goal (VTI), and verbs with animate actors and goal (VTA). For verbs, clustering took place within each of these proto-classes. Among the VIIs, 10 classes proved optimal, VAIs had 25 classes, VTIs with 15 classes, and VTAs with 20 classes. The choice to preprocess verbs into these four classes was as not doing so resulted in a clus-tering pattern that focused mainly on the difference between transitivity and the animacy of arguments. Any more or fewer classes and HAC clusters were far less cohesive with obvious semantic units being dispersed among many classes or split into multiple classes with no obvious differentiation. Similarly, verbs were split from nouns in this process because definitions in Wolvengrey (2001) vary significantly between verbs and nouns.",
                "cite_spans": [
                    {
                        "start": 1061,
                        "end": 1078,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "Nouns are naturally divided into two main classes in N\u00eahiyaw\u00eawin: animate and inanimate. 6 For our purposes we divide these further within each class between independent (i.e. alienable) and dependent (i.e. inalienable) nouns to create four main classes: Independent Animate Nouns (NA), Dependent Animate Nouns (NDA), Independent inanimate Nouns (NI), and Dependent Inanimate Nouns (NDI). The reason for this further division is due to the morphosemantic differences between independent and dependent nouns in N\u00eahiyaw\u00eawin. While independent nouns can stand on their own and represent a variety of entities, they are semantically and morphologically dependent on some possessor. We opted to pre-split NDIs and NDAs into their own classes, so as not to have the clustering focus on alienablity as the most major difference. 7",
                "cite_spans": [
                    {
                        "start": 89,
                        "end": 90,
                        "text": "6",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Methodology",
                "sec_num": "3"
            },
            {
                "text": "In all cases, clusters produced by this procedure needed some amount of post-processing. For nouns, this post-processing was minimal and mostly took the form of adjustments to the produced clusters: moving some items from one class to another, splitting a class that had clear semantic divisions, etc. For the verbs, this processing was often more complex, especially for the VAI and VTA classes. Items were determined to not belong in one class or another based on it's central meaning of the action or entity. If the majority of group members pertained to smoking (a cigarette), a word describing smokiing meat (as food preparation) would not be placed in this group, as the essence of the action and its intended purpose diverged significantly from the rest of the group. 6 Although this gender dichotomy is mostly semantically motivated (e.g. nouns that are semantically inanimate are part of the inanimate gender) this is not always the case as in the word pahkw\u00easikan, 'bread', a grammatically animate word.",
                "cite_spans": [
                    {
                        "start": 775,
                        "end": 776,
                        "text": "6",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "7 Preliminary results for words not seperated by their conjugation class or declension did, in fact, create clusters based around these obvious differences. This likely due to the way definitions were phrased (e.g. dependent nouns would have a possessive determiner or pronoun).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "Although most clusters produced somewhat cohesive semantic units, the largest clusters for the VAI and VTA classes acted as, essentially, catch-all clusters. Although computationally they seemed to have similar vector semantics, the relationship between items was not obvious to the human eye. Postprocessing for these clusters took substantial amounts of time and essentially comprised of using more cohesive clusters as a scaffold to fit words from these catch-all clusters into. In most cases, this resulted in slightly more clusters after postprocessing, though for VAIs this number was significantly higher, and for the NDIs it was slightly lower. Table 1 lists the number of cluster directly from HAC and from postprocessing.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 653,
                        "end": 660,
                        "text": "Table 1",
                        "ref_id": "TABREF1"
                    }
                ],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "Postprocessing grouped together words based on the most core semantic property of the word class: nouns were generally grouped based on the entity or state they represented, and verbs were generally grouped based on the most basic form action they represented. This is why, for example, AI-cover includes words for both covering and uncovering. In some cases a final class may seem like something that could be subsumed under another (e.g. AI-pray or AI-cooking might be understood as subsets of AI-action); however, in these cases, the subsumed class was judged to be sufficiently separate (e.g. cooking is an action of transforming resources into food for the purposes of nourishment, while verbs of AI-action are more manipulative, direct actions done for their own sake. Further, the automatic classification already grouped words in these ways, further justifying their separation. Finally, some grouping seem more morphosyntactic (e.g. AI-reflexive), though we argue that reflexivity, performing an action inwards, is in and of itself a salient semantic feature, and the inclusion of these terms into Wolvengrey (2001) indicates their lexicalization and distinction from the non-reflexive forms.",
                "cite_spans": [
                    {
                        "start": 1107,
                        "end": 1124,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "The actual quality of clustering varied form class to class. In general, nouns resulted in much more cohesive clusters out-of-the-box and required far less postprocessing. For example, nearly all of the HAC class N I 14 items referred to parts of human bodies (and those that did not fit this description were terms clearly related to body parts like aspat\u00e2skwahpisowin, 'back rest'), N I 13 was made up of trapping/hunting words and words for nests/animals. The NA classes produced through HAC were similarly straightforward: N I 9 was made up of words for trees, poles, sticks, and plants; N I 8 was made up entirely of words form beasts of burden, carts, wheels, etc.; while much of N A 3 and N A 7 , and nearly all of N A 2 referred to other animals. Once manually postprocessed, the NA lexemes settled into 8 classes: NA-persons, NA-beast-of-burden, NA-food, NA-celestial, NA-body-part, NA-religion, NA-money/count, and NA-shield.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "The NDI and NDA classes required almost no postprocessing: N DA 1 and N DA 3 were each made up of various family and non-family based relationships, while N DA 2 was made up of words for body parts and clothing. The resulting classes for these were: NDA-Relations, NDA-Body, and NDA-Clothing.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "The NDI lexemes basically took two classes: the vast majority of NDI forms referred to bodies and body parts while two lexemes referred to the concept of a house, resulting in only two classes: NDI-body, and NDI-house.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "Verbs, on the other hand, required quite a deal more postprocessing. VIIs showed the best clustering results without postprocessing. For example, V II 6 was entirely made up of taste/smell lexemes, V II 7 was almost entirely weather-related, V II 8 contained verbs that only take plural subjects, V II 9 had only lexemes referring to sound and sight, and V II 1 0 had only nominal-like verbs (e.g. m\u00eesiy\u00e2pisk\u00e2w '(it is) rust(y)'). Despite these well formed clusters, V II 1 through V II 5 were less cohesive and required manual clustering. In the end, distinct classes were identified: II-natural-land, II-weather-time, II-sensory-attitude, II-plural, II-move, II-time, and II-named. 8 Although postprocessing was required, this was not too substantial in scope or time.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "The VAIs required significantly more work. Some classes were well defined, such as V AI 23 whose members all described some sort of flight, but V AI 12 contains verbs of expectoration, singing, dancing, and even painting. The HAC classes were consolidated into 13 classes: AI-state, AI-action, AI-reflexive, AI-cooking, AI-speech, AI-collective, AI-care, AI-heat/fire, AI-money/count, AI-pray, AI-childcare, AI-canine, and AI-cover.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "The VTIs similarly required manual postprocessing after HAC clustering. Although some classes such as V T I 11 (entirely to do with cutting or breaking) or V T I 14 (entirely to do with pulling) were very well formed, the majority of the classes needed further subdivision (though significantly less so than with the VAIs, resulting in the following 6 classes: TI-action, TI-nonaction, TI-speech, TI-money/counter, TI-fit, and TI-food.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "Finally, the VTAs required a similar amount of postpreocessing as the VAIs. Although a few classes were well formed (such as V T A 4 which was entirely made up of verbs for 'causing' something), the vast majority of HAC classes contained two or more clear semantic groupings. Through manual postprocessing, the following set of classes were defined: VTA_allow, VTA_alter, VTA_body-position, VTA_care-for, VTA_cause, VTA_clothes, VTA_cognition, VTA_create, VTA_deceive, VTA_do, VTA_existential, VTA_food, VTA_hunt, VTA_miss/err, VTA_money, VTA_move, VTA_play, VTA_restrain, VTA_religious, VTA_seek, VTA_sense, VTA_speech, VTA_teach, VTA_tire, VTA_treat-a-way, VTA_(un)cover",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Results",
                "sec_num": "4"
            },
            {
                "text": "In addition the above evaluation in the description of the manual scrutiny and adjustment of HAC results, which is in and of itself an evaluation of the technique presented in this paper (with singlesubject experimentation proposed as a rapid path to data for less-resourced languages such as Vietnamese (Pham and Baayen, 2015) ), we present a preliminary quantitative evaluation of this technique. This evaluation allows us to judge how useful these classes are in practical terms, providing an indirect measure of the informational value of the clusters. We make use of the mixed effects modelling that initially motivated automatic semantic clustering, focusing on a morphological alternation called N\u00eahiyaw\u00eawin Order, wherein a verb may take the form ninip\u00e2n (the Independent) or \u00ea-nip\u00e2y\u00e2n (the \u00ea-Conjunct), both of which may be translated as 'I sleep.' The exact details of this alternation remain unclear, though there appears to be some syntactic and pragmatic motivation (Cook, 2014) . Using R (R Core Team, 2017) and the lme4 package (Bates et al., 2015) , we ran a logistic regression to predict alternation using verbal semantic classes as categorical variables. In order to isolate the effect of semantic class, no other effects were used. The semantic classes were included as random effects. To assess the effectiveness of semantic class in this context, we assess the pseudo-R 2 value, a measure of Goodness-of-Fit. Unlike a regular R 2 measure, the pseudo-R 2 can not be interpreted as a direct measure of how much a model explains variance, and generally \"good\" pseudo-R 2 value are comparatively smaller (McFadden et al., 1973) , though a higher value still represents a better fit. As a general rule, a pseudo-R 2 of 0.20 to 0.40 represents a well fit model. 1977) 9 Models were fit for each of the four conjugation classes for both classes produced directly from the Hierarchical Agglomerative Clustering as well those manually adjusted. We used a subset of the Ahenakew-Wolfart Corpus (Arppe et al., 2020) , containing 10,764 verb tokens observed in either the Independent or \u00ea-Conjunct forms. The resulting pseudo-R 2 scores represent the way in which automatic and semi-manual clusters can explain the N\u00eahiyaw\u00eawin Order alternation. Table 2 presents the result of these analyses. the Manual column represents clusters that were manually adjusted, while the HAC-Only column represents the result of the logistic model that used only the fully automatic HAC-produced clusters. The manually adjusted and HAC-only classes performed similarly, especially for VTAs, though manual adjustment had a slightly worse fit for the VIIs, and conversely the VAI and VTI has somewhat significantly better fits using the manually adjusted classes. Although it appears that manual adjustment produced classes that were somewhat better able to explain this alternation, both manually adjusted and HAC-only clusters appear to explain a non-negligible degree of this alternation phenomenon in the above models. This is significant, because it shows that the result of the clustering techniques presented in this paper produce a tangible and useful product for linguistic analysis. Further, it suggests that, although manual classification was sometimes more useful, automatic classes more or less performed as well, allowing for researchers to determine if the added effort is worth the small increase in informational value. Nevertheless, alternative methods of evaluation, such as evaluating clusters based on speaker input, particularly through visual meas as described in Majewska et al. (2020) should be considered. 10 9 One can also compare the results in this paper with results from a similar alternation study in Arppe et al. (2008) .",
                "cite_spans": [
                    {
                        "start": 304,
                        "end": 327,
                        "text": "(Pham and Baayen, 2015)",
                        "ref_id": "BIBREF16"
                    },
                    {
                        "start": 979,
                        "end": 991,
                        "text": "(Cook, 2014)",
                        "ref_id": "BIBREF4"
                    },
                    {
                        "start": 1043,
                        "end": 1063,
                        "text": "(Bates et al., 2015)",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 1622,
                        "end": 1645,
                        "text": "(McFadden et al., 1973)",
                        "ref_id": "BIBREF12"
                    },
                    {
                        "start": 2006,
                        "end": 2026,
                        "text": "(Arppe et al., 2020)",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 3578,
                        "end": 3600,
                        "text": "Majewska et al. (2020)",
                        "ref_id": "BIBREF10"
                    },
                    {
                        "start": 3724,
                        "end": 3743,
                        "text": "Arppe et al. (2008)",
                        "ref_id": "BIBREF1"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 2256,
                        "end": 2263,
                        "text": "Table 2",
                        "ref_id": "TABREF3"
                    }
                ],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "4.1"
            },
            {
                "text": "10 It is worth noting that previous attempts at such experi-",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "4.1"
            },
            {
                "text": "In general, the best clustering was seen in classes with fewer items. The VAI and NI lexemes required the most postprocessing, with each having roughly double the number of items as the next most numerous verb/noun class. Verb classes in general seemed to produce less cohesive classes through HAC. Although the exact cause of this discrepancy in unknown, it could perhaps be due to the way words are defined in Wolvengrey (2001) . In this dictionary, verb definitions almost always contain more words than noun definitions. Almost every single verb definition will have at least two words, owing to the fact that N\u00eahiyaw\u00eawin verbs are defined by an inflected lexeme. This means that if one looks up a word like walk, it would appear as: pimoht\u00eaw: s/he walks, s/he walks along; s/he goes along. Meanwhile, nouns tend to have shorter definitions. The definition for the act of walking, a nominalized form of the verb for walk, is written as: pimoht\u00eawin: walk, stroll; sidewalk. This difference is exacerbated by the fact that definitions are often translated fairly literally. Something like p\u00eayakw\u00eayimisow might be translated simply as 's/he is selfish,' but contains morphemes meaning one, think, reflexive, and s/he. A gloss of this word is seen in (1). Rather than simply defining the word as 's/he is selfish,' (Wolvengrey, 2001) has opted to provide a more nuanced definition: p\u00eayakw\u00eayimisow: s/he thinks only of him/herself, s/he is selfish, s/he is self-centered.",
                "cite_spans": [
                    {
                        "start": 412,
                        "end": 429,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 1315,
                        "end": 1333,
                        "text": "(Wolvengrey, 2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "(1) p\u00eayakw\u00eayimisow p\u00eayakw-\u00eayi-m-iso-w one-think-VTA-RFLX-3SG 's/he thinks only of him/herself' The result of this complex form of defining is that words are defined more in line with how they are understood within the N\u00eahiyaw\u00eawin culture, which is indeed often manifested in the derivational morphological composition of these words. This is central to the motivation for this method of semiautomatic clustering, but produces verbs with relatively long definitions. An alternative explanation for why N\u00eahiyaw\u00eawin lexemes with English definitions consisting of more numerous parts of speech were more difficult to classify is that these divisions simply have significantly more variation in mentation via N\u00eahiyaw\u00eawin communities with which we have good relationships have been poorly received by speakers. meaning for whatever reason. Further investigation into this is needed.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "Also worth noting is the relative distributions of each of the postprocessed classes mentioned above. Table 3 details each of the postprocessed noun classes sorted by their size.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 102,
                        "end": 109,
                        "text": "Table 3",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "Perhaps unsurprisingly, the distribution of lexemes into different classes followed a sort of Zipfian distribution.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "The NA-person and NA-other-animals accounted for the vast majority of noun lexemes for animate nouns. Just under half of all NI lexemes were nominalized verbs, and roughly a quarter were smaller objectlike items (e.g. tools, dishes, etc.). The NDAs were almost entirely dominated by words for family, while all but three NDIs were body part lexemes. Some categories such as NI-scent, NI-days, and NA-shield have extremely low membership counts, but were substantially different from other categories that they were not grouped into another class. Most interestingly, there appeared to be three NI lexemes that referred to persons, something usually reserved for NAs only. These lexemes were okitaham\u00e2k\u00eaw 'one who forbids,' owiyasiw\u00eawikim\u00e2w 'magistrate,' and mihkokwayaw\u00eaw 'red neck.' In all three cases, the lexemes seem to be deverbal nouns (from kitaham\u00e2k\u00eaw 's/he forbids,' wiyasiw\u00eaw 's/he makes laws,' and mihkokwayaw\u00eaw 's/he has a red neck.'",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "Verbs showed a similar distribution. Table 4 details the distribution of words within each of semantic classes for verbs. With the exception of VII and VAIs, verbs were dominated by classes for action, which subsumes most volitional actions (e.g. k\u00eeskihkw\u00eapisiw\u00eaw 's/he rips the face off of people,' k\u00e2s\u00eepayiw 's/he deletes'), and nonaction which includes most verbs of thought, emotion, judgment, or sensory action (e.g koskowih\u00eaw, 's/he startles someone,' n\u00f4c\u00eehkaw\u00eaw 's/he seduces someone'). Other classes may include action verbs, such as AI-cooking and TI-speech. Although these verbs could be classified in one of the two previously mentioned systems, their automatic classification and semantics unify them in a way that is unique to other items in these larger classes.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 37,
                        "end": 44,
                        "text": "Table 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "Overall, verb forms, especially the most numerous classes of VAI and VTA, required a large degree of manual postprocessing. Because this approach assumes no underlying ontology, but rather attempts to work bottom-up (cf. Hanks (1996) ), the time taken to postprocess VAI and VTA classes is likely not too far from what it would take to manually classify these words based off a prebuilt ontology; however, the appeal of a bottom-up classification should not be overlooked, however. As an example, many ontologies place concepts like thinking, and being happy into separate classes; however, in our classification these words were combined into a single class of cognition. This is done because emotion words like m\u00f4cik\u00eayihtam, 's/he is happy (because of something)' (in addition to being verbs and not adjectives) contain a morpheme, {-\u00eayi-}, meaning 'thought.' For these reasons, such emotion words are often translated as having to do specifically with thought and cognition: m\u00f4cik\u00eayihtam, 's/he thinks happily (because of something).' (Wolvengrey, 2001 ) uses these sorts of definitions, and so unsurprisingly the majority of such emotion words were classified in the proposed scheme together with words of thought. Where this was not the case, manual postprocessing from a bottom-up approach allows us to maintain the cultural understanding of emotions as directly related to cognition. Furthermore, from the experiential standpoint of one of the authors, the use of semi-automatic clustering produces a kick-start that greatly aids to the starting of a semantic classification task, especially for non-native speakers.",
                "cite_spans": [
                    {
                        "start": 221,
                        "end": 233,
                        "text": "Hanks (1996)",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 1038,
                        "end": 1055,
                        "text": "(Wolvengrey, 2001",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Discussion",
                "sec_num": "5"
            },
            {
                "text": "This paper describes an attempt at, for the first time, semi-automatically classifying N\u00eahiyaw\u00eawin verbs and nouns. The process used in this paper is easily applied to any language that makes use of a bilingual dictionary with definitions written in a more resourced language. Resulting clusters of N\u00eahiyaw\u00eawin words are freely available on the online. Although the technique worked quite well with nouns, which required very little manual adjustment, verbs required more directed attention. Despite this, the technique presented in this paper offers a bottom-up, data-driven approach that takes the language on its own terms, without resorting to ontologies created primarily for other languages. If, however, one wishes to use a pre-defined ontology, the basis for this work (representing word definitions using pre-trained English word vectors) could be used in conjunction with existing ontologies to expedite the classification process. For example, Dacanay et al. (2021) compare the naive definition vectors for Wolvengrey (2001) with the same for the English WordNet word senses; word senses",
                "cite_spans": [
                    {
                        "start": 955,
                        "end": 976,
                        "text": "Dacanay et al. (2021)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 1018,
                        "end": 1035,
                        "text": "Wolvengrey (2001)",
                        "ref_id": "BIBREF20"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "6"
            },
            {
                "text": "See http://rapidwords.net/",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "This corpus was trained on a large corpus of 100 billion words. Available at https://code.google.com/ archive/p/word2vec/",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "In reality, there were only a handful of cases where words occurred in the dictionary but not in the Google News Corpus. Because there are so few examples of this, even simply leaving these items out would not substantiqally change clustering results.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "As discussed inWolvengrey (2005), N\u00eahiyaw\u00eawin sentences are devoid subject and objects in the usual sense. Instead, syntactic roles are defined by verbal direction alignment. For this reason, we use the terms actor and goal instead of subject and object.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The concepts of weather and time were combined here as many of the N\u00eahiyaw\u00eawin words for specific times also contain some concept of weather (e.g. the term for 'day' is k\u00eesik\u00e2w, clearly related to the word for 'sky/heavens', k\u00eesik; similarly, the word for 'night' is tipisk\u00e2w, which is the same word used for the night sky. Additionally, words like pipon, 'winter' and s\u00eekwan 'spring' are obviously related to both time and weather.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [
            {
                "text": "NDI (N) NA (N) NDA (N) NI-nominal (1783) NDI-body (243) NA-persons (720) NDA-relations (143) NI-object 902NDI-house (2) NA-beast-of-burden (512) NDA-body (45) NI-natural-Force 283NA-food (325) NDA-clothing (4) NI-place 228NA-celestial (45) NI-nature-plants 198NA-body-part (37) NI-body-part 78NA-religion (23) NI-hunt-trap 60NA-money/count (12) NI-animal-product 48NA-shield (2) NI-religion (36) NI-alteration (23) NI-scent (4) NI-days (4) NI-persons (3)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "NI (N)",
                "sec_num": null
            },
            {
                "text": "We would like to thank the Social Sciences and Humanities Research Council of Canada for funding this research. We would also like to thank Dr. Arok Wolvengrey for providing his dictionary source for this study.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Acknowledgements",
                "sec_num": null
            },
            {
                "text": " 35AI-speech (131) TI-fit 10TA-money/count (23) II-named 3AI-collective 97TI-food 8TA-religion (9) AI-care 81TA-allow (5) AI-heat/fire (55) AI-money/count (34) AI-pray (29) AI-childcare (17) AI-canine (16) AI-cover (15) whose vectors bear a strong correlation with the N\u00eahiyaw\u00eawin definitions can then be assumed to be semantically similar with a N\u00eahiyaw\u00eawin word, and the latter can take the WordNet classification of the former. Further research should investigate more sophisticated methods of creating embeddings, especially the use of true sentence vectors. Additionally, one could consider using weights for English words in the definitions of N\u00eahiyaw\u00eawin words based on measures like tf-idf. Over all, this technique provided promising results. Regardless of the language or particular implementation, this technique of bootstrapping under-resourced language data with pre-trained majority language vectors (for which very large corpora exist), should not be restricted by the sizes of dictionaries in the under-resourced language, as the underlying vectors are trained on a 100 million word English corpus.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "annex",
                "sec_num": null
            }
        ],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "A morphosyntactically tagged corpus for plains cree",
                "authors": [
                    {
                        "first": "Antti",
                        "middle": [],
                        "last": "Arppe",
                        "suffix": ""
                    },
                    {
                        "first": "Katherine",
                        "middle": [],
                        "last": "Schmirler",
                        "suffix": ""
                    },
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Atticus",
                        "suffix": ""
                    },
                    {
                        "first": "Arok",
                        "middle": [],
                        "last": "Harrigan",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Wolvengrey",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "Papers of the Forty-Ninth Algonquian Conference",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Antti Arppe, Katherine Schmirler, Atticus G Harrigan, and Arok Wolvengrey. 2020. A morphosyntactically tagged corpus for plains cree. In Papers of the Forty- Ninth Algonquian Conference. Michigan State Uni- versity Press.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Univariate, bivariate, and multivariate methods in corpus-based lexicography: A study of synonymy",
                "authors": [
                    {
                        "first": "Antti",
                        "middle": [],
                        "last": "Arppe",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Antti Arppe et al. 2008. Univariate, bivariate, and mul- tivariate methods in corpus-based lexicography: A study of synonymy.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Fitting linear mixed-effects models using lme4",
                "authors": [
                    {
                        "first": "Douglas",
                        "middle": [],
                        "last": "Bates",
                        "suffix": ""
                    },
                    {
                        "first": "Martin",
                        "middle": [],
                        "last": "M\u00e4chler",
                        "suffix": ""
                    },
                    {
                        "first": "Ben",
                        "middle": [],
                        "last": "Bolker",
                        "suffix": ""
                    },
                    {
                        "first": "Steve",
                        "middle": [],
                        "last": "Walker",
                        "suffix": ""
                    }
                ],
                "year": 2015,
                "venue": "Journal of Statistical Software",
                "volume": "67",
                "issue": "1",
                "pages": "1--48",
                "other_ids": {
                    "DOI": [
                        "10.18637/jss.v067.i01"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Douglas Bates, Martin M\u00e4chler, Ben Bolker, and Steve Walker. 2015. Fitting linear mixed-effects models using lme4. Journal of Statistical Software, 67(1):1- 48.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Predicting the dative alternation",
                "authors": [
                    {
                        "first": "Joan",
                        "middle": [],
                        "last": "Bresnan",
                        "suffix": ""
                    },
                    {
                        "first": "Anna",
                        "middle": [],
                        "last": "Cueni",
                        "suffix": ""
                    },
                    {
                        "first": "Tatiana",
                        "middle": [],
                        "last": "Nikitina",
                        "suffix": ""
                    },
                    {
                        "first": "R Harald",
                        "middle": [],
                        "last": "Baayen",
                        "suffix": ""
                    }
                ],
                "year": 2007,
                "venue": "Cognitive foundations of interpretation",
                "volume": "",
                "issue": "",
                "pages": "69--94",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Joan Bresnan, Anna Cueni, Tatiana Nikitina, and R Harald Baayen. 2007. Predicting the dative alter- nation. In Cognitive foundations of interpretation, pages 69-94. KNAW.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "The clause-typing system of Plains Cree: Indexicality, anaphoricity, and contrast",
                "authors": [
                    {
                        "first": "Clare",
                        "middle": [],
                        "last": "Cook",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Clare Cook. 2014. The clause-typing system of Plains Cree: Indexicality, anaphoricity, and contrast, vol- ume 2. OUP Oxford.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Computational Analysis versus Human Intuition: A Critical Comparison of Vector Semantics with Manual Semantic Classification in the Context of Plains Cree",
                "authors": [
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Dacanay",
                        "suffix": ""
                    },
                    {
                        "first": "Antti",
                        "middle": [],
                        "last": "Arppe",
                        "suffix": ""
                    },
                    {
                        "first": "Atticus",
                        "middle": [],
                        "last": "Harrigan",
                        "suffix": ""
                    }
                ],
                "year": 2021,
                "venue": "Proceedings of the 4th Workshop on the Use of Computational Methods in the Study of Endangered Languages",
                "volume": "1",
                "issue": "",
                "pages": "33--43",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Daniel Dacanay, Antti Arppe, and Atticus Harrigan. 2021. Computational Analysis versus Human Intu- ition: A Critical Comparison of Vector Semantics with Manual Semantic Classification in the Context of Plains Cree. In Proceedings of the 4th Workshop on the Use of Computational Methods in the Study of Endangered Languages, volume 1, pages 33-43.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Ways of trying in russian: Clustering behavioral profiles. Corpus linguistics and linguistic theory",
                "authors": [
                    {
                        "first": "Dagmar",
                        "middle": [],
                        "last": "Divjak",
                        "suffix": ""
                    },
                    {
                        "first": "Stefan",
                        "middle": [
                            "Th"
                        ],
                        "last": "Gries",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "",
                "volume": "2",
                "issue": "",
                "pages": "23--60",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Dagmar Divjak and Stefan Th Gries. 2006. Ways of try- ing in russian: Clustering behavioral profiles. Cor- pus linguistics and linguistic theory, 2(1):23-60.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Automatic information extraction and building of lexical semantic resources for NLP applications",
                "authors": [
                    {
                        "first": "Birgit",
                        "middle": [],
                        "last": "Hamp",
                        "suffix": ""
                    },
                    {
                        "first": "Helmut",
                        "middle": [],
                        "last": "Feldweg",
                        "suffix": ""
                    }
                ],
                "year": 1997,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Birgit Hamp and Helmut Feldweg. 1997. Germanet-a lexical-semantic net for german. In Automatic infor- mation extraction and building of lexical semantic resources for NLP applications.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Contextual dependency and lexical sets",
                "authors": [
                    {
                        "first": "Patrick",
                        "middle": [],
                        "last": "Hanks",
                        "suffix": ""
                    }
                ],
                "year": 1996,
                "venue": "International journal of corpus linguistics",
                "volume": "1",
                "issue": "1",
                "pages": "75--98",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Patrick Hanks. 1996. Contextual dependency and lexi- cal sets. International journal of corpus linguistics, 1(1):75-98.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "English verb classes and alternations: A preliminary investigation",
                "authors": [
                    {
                        "first": "Beth",
                        "middle": [],
                        "last": "Levin",
                        "suffix": ""
                    }
                ],
                "year": 1993,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Beth Levin. 1993. English verb classes and alterna- tions: A preliminary investigation. University of Chicago press.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Manual clustering and spatial arrangement of verbs for multilingual evaluation and typology analysis",
                "authors": [
                    {
                        "first": "Olga",
                        "middle": [],
                        "last": "Majewska",
                        "suffix": ""
                    },
                    {
                        "first": "Ivan",
                        "middle": [],
                        "last": "Vuli\u0107",
                        "suffix": ""
                    },
                    {
                        "first": "Diana",
                        "middle": [],
                        "last": "Mccarthy",
                        "suffix": ""
                    },
                    {
                        "first": "Anna",
                        "middle": [],
                        "last": "Korhonen",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "Proceedings of the 28th International Conference on Computational Linguistics",
                "volume": "",
                "issue": "",
                "pages": "4810--4824",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Olga Majewska, Ivan Vuli\u0107, Diana McCarthy, and Anna Korhonen. 2020. Manual clustering and spa- tial arrangement of verbs for multilingual evaluation and typology analysis. In Proceedings of the 28th International Conference on Computational Linguis- tics, pages 4810-4824.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Quantitative Methods for Analyzing Travel Behaviour of Individuals: Some Recent Developments",
                "authors": [
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Mcfadden",
                        "suffix": ""
                    }
                ],
                "year": 1977,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Daniel McFadden. 1977. Quantitative Methods for An- alyzing Travel Behaviour of Individuals: Some Re- cent Developments. Technical report.",
                "links": null
            },
            "BIBREF12": {
                "ref_id": "b12",
                "title": "Conditional logit analysis of qualitative choice behavior",
                "authors": [
                    {
                        "first": "Daniel",
                        "middle": [],
                        "last": "Mcfadden",
                        "suffix": ""
                    }
                ],
                "year": 1973,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Daniel McFadden et al. 1973. Conditional logit analy- sis of qualitative choice behavior.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "Efficient estimation of word representations in vector space",
                "authors": [
                    {
                        "first": "Tomas",
                        "middle": [],
                        "last": "Mikolov",
                        "suffix": ""
                    },
                    {
                        "first": "Kai",
                        "middle": [],
                        "last": "Chen",
                        "suffix": ""
                    },
                    {
                        "first": "Greg",
                        "middle": [],
                        "last": "Corrado",
                        "suffix": ""
                    },
                    {
                        "first": "Jeffrey",
                        "middle": [],
                        "last": "Dean",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {
                    "arXiv": [
                        "arXiv:1301.3781"
                    ]
                },
                "num": null,
                "urls": [],
                "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jef- frey Dean. 2013a. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Distributed representations of words and phrases and their compositionality",
                "authors": [
                    {
                        "first": "Tomas",
                        "middle": [],
                        "last": "Mikolov",
                        "suffix": ""
                    },
                    {
                        "first": "Ilya",
                        "middle": [],
                        "last": "Sutskever",
                        "suffix": ""
                    },
                    {
                        "first": "Kai",
                        "middle": [],
                        "last": "Chen",
                        "suffix": ""
                    },
                    {
                        "first": "Greg",
                        "middle": [
                            "S"
                        ],
                        "last": "Corrado",
                        "suffix": ""
                    },
                    {
                        "first": "Jeff",
                        "middle": [],
                        "last": "Dean",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Advances in neural information processing systems",
                "volume": "",
                "issue": "",
                "pages": "3111--3119",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- rado, and Jeff Dean. 2013b. Distributed representa- tions of words and phrases and their compositional- ity. In Advances in neural information processing systems, pages 3111-3119.",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "WordNet: An electronic lexical database",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "George",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "Miller",
                        "suffix": ""
                    }
                ],
                "year": 1998,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "George A Miller. 1998. WordNet: An electronic lexical database. MIT press.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "Vietnamese compounds show an anti-frequency effect in visual lexical decision. Language, Cognition and Neuroscience",
                "authors": [
                    {
                        "first": "Hien",
                        "middle": [],
                        "last": "Pham",
                        "suffix": ""
                    },
                    {
                        "first": "Harald",
                        "middle": [],
                        "last": "Baayen",
                        "suffix": ""
                    }
                ],
                "year": 2015,
                "venue": "",
                "volume": "30",
                "issue": "",
                "pages": "1077--1095",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hien Pham and Harald Baayen. 2015. Vietnamese compounds show an anti-frequency effect in visual lexical decision. Language, Cognition and Neuro- science, 30(9):1077-1095.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing",
                "authors": [
                    {
                        "first": "",
                        "middle": [],
                        "last": "R Core Team",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "R Core Team. 2017. R: A Language and Environment for Statistical Computing. R Foundation for Statisti- cal Computing, Vienna, Austria.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "Eastern james bay cree thematic dictionary (southern dialect)",
                "authors": [
                    {
                        "first": "Linda",
                        "middle": [],
                        "last": "Visitor",
                        "suffix": ""
                    },
                    {
                        "first": "Marie-Odile",
                        "middle": [],
                        "last": "Junker",
                        "suffix": ""
                    },
                    {
                        "first": "Mimie",
                        "middle": [],
                        "last": "Neacappo",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Linda Visitor, Marie-Odile Junker, and Mimie Nea- cappo. 2013. Eastern james bay cree thematic dic- tionary (southern dialect). Chisasibi: Cree School Board.",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "Building the chinese open wordnet (cow): Starting from core synsets",
                "authors": [
                    {
                        "first": "Shan",
                        "middle": [],
                        "last": "Wang",
                        "suffix": ""
                    },
                    {
                        "first": "Francis",
                        "middle": [],
                        "last": "Bond",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Proceedings of the 11th Workshop on Asian Language Resources",
                "volume": "",
                "issue": "",
                "pages": "10--18",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Shan Wang and Francis Bond. 2013. Building the chi- nese open wordnet (cow): Starting from core synsets. In Proceedings of the 11th Workshop on Asian Lan- guage Resources, pages 10-18.",
                "links": null
            },
            "BIBREF20": {
                "ref_id": "b20",
                "title": "N\u0113hiyaw\u0113win: itw\u0113wina = Cree: words",
                "authors": [
                    {
                        "first": "Arok",
                        "middle": [],
                        "last": "Wolvengrey",
                        "suffix": ""
                    }
                ],
                "year": 2001,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Arok Wolvengrey. 2001. N\u0113hiyaw\u0113win: itw\u0113wina = Cree: words. University of Regina press.",
                "links": null
            },
            "BIBREF21": {
                "ref_id": "b21",
                "title": "Inversion and the absence of grammatical relations in plains cree. Morphosyntactic expression in functional grammar",
                "authors": [
                    {
                        "first": "Arok",
                        "middle": [],
                        "last": "Wolvengrey",
                        "suffix": ""
                    }
                ],
                "year": 2005,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Arok Wolvengrey. 2005. Inversion and the absence of grammatical relations in plains cree. Morphosyntac- tic expression in functional grammar, 27.",
                "links": null
            },
            "BIBREF22": {
                "ref_id": "b22",
                "title": "Refining word embeddings for sentiment analysis",
                "authors": [
                    {
                        "first": "Liang-Chih",
                        "middle": [],
                        "last": "Yu",
                        "suffix": ""
                    },
                    {
                        "first": "Jin",
                        "middle": [],
                        "last": "Wang",
                        "suffix": ""
                    },
                    {
                        "first": "Robert",
                        "middle": [],
                        "last": "Lai",
                        "suffix": ""
                    },
                    {
                        "first": "Xuejie",
                        "middle": [],
                        "last": "Zhang",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "Proceedings of the 2017 conference on empirical methods in natural language processing",
                "volume": "",
                "issue": "",
                "pages": "534--539",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Liang-Chih Yu, Jin Wang, K Robert Lai, and Xuejie Zhang. 2017. Refining word embeddings for senti- ment analysis. In Proceedings of the 2017 confer- ence on empirical methods in natural language pro- cessing, pages 534-539.",
                "links": null
            }
        },
        "ref_entries": {
            "TABREF1": {
                "num": null,
                "type_str": "table",
                "text": "HAC built cluster counts vs. counts after postprocessing",
                "content": "<table/>",
                "html": null
            },
            "TABREF3": {
                "num": null,
                "type_str": "table",
                "text": "",
                "content": "<table><tr><td>: pseudo-R 2 Values for Modelling Independent</td></tr><tr><td>vs. \u00ea-Conjunct Order Choice Based on Manual and Au-</td></tr><tr><td>tomatic Clustering Evaluation</td></tr></table>",
                "html": null
            }
        }
    }
}