mitochondriaext commited on
Commit
1ec0c39
·
verified ·
1 Parent(s): c096112

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: jcblaise/gpt2-tagalog
3
+ datasets:
4
+ - jcblaise/hatespeech_filipino
5
+ language:
6
+ - tl
7
+ - en
8
+ ---
9
+
10
+ # Model Card for Model ID
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+ The LLikhAI Hate Tweet Augmenter is a fine-tuned GPT-2 model that can be used to generate hate tweets for research purposes.
15
+
16
+ ## Model Details
17
+
18
+ ### Model Description
19
+
20
+ <!-- Provide a longer summary of what this model is. -->
21
+
22
+ LLikhAI Hate Tweet Augmenter a GPT-2 model fine-tuned using the hate instances of the Hate Speech Dataset made by Blaise-Cruz and Cheng (2019). The model is made to augment low-resource Filipino hate speech datasets in order for these to have more instances in the hopes of developing more robust models.
23
+
24
+ - **Language(s) (NLP):** All languages from original GPT-2 model. Fine-tuned for Tagalog.
25
+ - **Finetuned from model:** jcblaise/gpt2-tagalog
26
+
27
+ ## Uses
28
+
29
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
30
+
31
+ ### Direct Use
32
+
33
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
34
+
35
+ To use the model, simply load the model from HuggingFace to your environment and input a chat-based prompt to allow the model to generate a new hate tweet.
36
+
37
+ ### Out-of-Scope Use
38
+
39
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
40
+
41
+ This model is not supposed to be used to spread hate speech, misinformation, violence, etc. on social media. This model is only supposed to be used for research purposes and for augmenting hate speech datasets.
42
+
43
+ ## Bias, Risks, and Limitations
44
+
45
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
46
+
47
+ This is GPT-2 model finetuned on the Hate Speech Dataset made by Blaise-Cruz and Cheng (2019). As such, this model carries the scope and limitations of both the original model and the original dataset. Given that the Hate Speech Dataset contains mostly political tweets, the generated tweets will have political tones.