cbdb commited on
Commit
ff7aaa0
·
1 Parent(s): e9c93b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -10,7 +10,7 @@ tags:
10
  license: cc-by-nc-sa-4.0
11
  ---
12
 
13
- # <font color="IndianRed"> BertForSequenceClassification model (Classical Chinese) </font>
14
  [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1aIGvyvqRdHv7QTRahhD1sf8L6yV39kxc?usp=sharing/)
15
 
16
  The Kraft (Korean Romanization From Transformer) model translates the characters (Hangul) of a Korean person name into the Roman alphabet ([McCune–Reischauer system](https://en.wikipedia.org/wiki/McCune%E2%80%93Reischauer)). Kraft uses the Transformer architecture, which is a type of neural network architecture that was introduced in the 2017 paper "Attention Is All You Need" by Google researchers. It is designed for sequence-to-sequence tasks, such as machine translation, language modeling, and summarization.
 
10
  license: cc-by-nc-sa-4.0
11
  ---
12
 
13
+ # <font color="IndianRed"> Kraft (Korean Romanization From Transformer) </font>
14
  [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1aIGvyvqRdHv7QTRahhD1sf8L6yV39kxc?usp=sharing/)
15
 
16
  The Kraft (Korean Romanization From Transformer) model translates the characters (Hangul) of a Korean person name into the Roman alphabet ([McCune–Reischauer system](https://en.wikipedia.org/wiki/McCune%E2%80%93Reischauer)). Kraft uses the Transformer architecture, which is a type of neural network architecture that was introduced in the 2017 paper "Attention Is All You Need" by Google researchers. It is designed for sequence-to-sequence tasks, such as machine translation, language modeling, and summarization.