dkawahara commited on
Commit
40948ff
·
1 Parent(s): 07adfbf

Updated README.md.

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -9,7 +9,7 @@ widget:
9
  - text: "早稲田 大学 で 自然 言語 処理 を [MASK] する 。"
10
  ---
11
 
12
- # nlp-waseda/roberta-large-japanese
13
 
14
  ## Model description
15
 
@@ -20,8 +20,8 @@ This is a Japanese RoBERTa large model pretrained on Japanese Wikipedia and the
20
  You can use this model for masked language modeling as follows:
21
  ```python
22
  from transformers import AutoTokenizer, AutoModelForMaskedLM
23
- tokenizer = AutoTokenizer.from_pretrained("nlp-waseda/roberta-large-japanese")
24
- model = AutoModelForMaskedLM.from_pretrained("nlp-waseda/roberta-large-japanese")
25
 
26
  sentence = '早稲田 大学 で 自然 言語 処理 を [MASK] する 。' # input should be segmented into words by Juman++ in advance
27
  encoding = tokenizer(sentence, return_tensors='pt')
 
9
  - text: "早稲田 大学 で 自然 言語 処理 を [MASK] する 。"
10
  ---
11
 
12
+ # nlp-waseda/roberta-large-japanese-seq512
13
 
14
  ## Model description
15
 
 
20
  You can use this model for masked language modeling as follows:
21
  ```python
22
  from transformers import AutoTokenizer, AutoModelForMaskedLM
23
+ tokenizer = AutoTokenizer.from_pretrained("nlp-waseda/roberta-large-japanese-seq512")
24
+ model = AutoModelForMaskedLM.from_pretrained("nlp-waseda/roberta-large-japanese-seq512")
25
 
26
  sentence = '早稲田 大学 で 自然 言語 処理 を [MASK] する 。' # input should be segmented into words by Juman++ in advance
27
  encoding = tokenizer(sentence, return_tensors='pt')