KoichiYasuoka
commited on
Commit
•
3e182db
1
Parent(s):
faba6e7
TransformersSlowUD
Browse files
README.md
CHANGED
@@ -59,7 +59,7 @@ class TaggerPipeline(TokenClassificationPipeline):
|
|
59 |
e=([x for x in a2b[i+1:] if x>[]]+[[len(text)]])[0][0]
|
60 |
d[i]["start"],d[i]["end"]=s,e
|
61 |
return d
|
62 |
-
class
|
63 |
def __init__(self,bert):
|
64 |
import os
|
65 |
self.tokenizer=AutoTokenizer.from_pretrained(bert)
|
@@ -105,7 +105,7 @@ class TransformersUD(object):
|
|
105 |
str(h[i]),p,"_","_" if i<n and w[i][0]<e else "SpaceAfter=No"])+"\n"
|
106 |
return u+"\n"
|
107 |
|
108 |
-
nlp=
|
109 |
print(nlp("全学年にわたって小学校の国語の教科書に挿し絵が用いられている"))
|
110 |
```
|
111 |
|
|
|
59 |
e=([x for x in a2b[i+1:] if x>[]]+[[len(text)]])[0][0]
|
60 |
d[i]["start"],d[i]["end"]=s,e
|
61 |
return d
|
62 |
+
class TransformersSlowUD(object):
|
63 |
def __init__(self,bert):
|
64 |
import os
|
65 |
self.tokenizer=AutoTokenizer.from_pretrained(bert)
|
|
|
105 |
str(h[i]),p,"_","_" if i<n and w[i][0]<e else "SpaceAfter=No"])+"\n"
|
106 |
return u+"\n"
|
107 |
|
108 |
+
nlp=TransformersSlowUD("KoichiYasuoka/deberta-large-japanese-unidic-ud-head")
|
109 |
print(nlp("全学年にわたって小学校の国語の教科書に挿し絵が用いられている"))
|
110 |
```
|
111 |
|