Preeti Dave commited on
Commit
fd3d614
·
1 Parent(s): 10e60a2
Files changed (5) hide show
  1. README.md +20 -0
  2. config.json +13 -0
  3. pytorch_model.bin +9 -0
  4. special_tokens_map.json +6 -0
  5. tokenizer.json +4 -0
README.md ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # My Custom BERT Model
2
+
3
+ This is a fine-tuned version of `bert-base-uncased` for binary sentiment analysis.
4
+
5
+ ## Model Overview
6
+ - Model: BERT for sequence classification
7
+ - Number of Labels: 2
8
+ - Dataset: Custom sentiment dataset
9
+
10
+ ## How to Use
11
+ ```python
12
+ from transformers import AutoModelForSequenceClassification, AutoTokenizer
13
+
14
+ # Load model and tokenizer
15
+ model = AutoModelForSequenceClassification.from_pretrained("your_username/my_model")
16
+ tokenizer = AutoTokenizer.from_pretrained("your_username/my_model")
17
+
18
+ # Tokenize text
19
+ inputs = tokenizer("I love this!", return_tensors="pt")
20
+ outputs = model(**inputs)
config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": ["BertForSequenceClassification"],
3
+ "hidden_size": 768,
4
+ "num_attention_heads": 12,
5
+ "num_hidden_layers": 12,
6
+ "vocab_size": 30522,
7
+ "type_vocab_size": 2,
8
+ "max_position_embeddings": 512,
9
+ "initializer_range": 0.02,
10
+ "layer_norm_eps": 1e-12,
11
+ "hidden_dropout_prob": 0.1,
12
+ "attention_probs_dropout_prob": 0.1
13
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import BertForSequenceClassification, BertTokenizer
2
+
3
+ # Initialize model and tokenizer
4
+ model = BertForSequenceClassification.from_pretrained("bert-base-uncased", num_labels=2)
5
+ tokenizer = BertTokenizer.from_pretrained("bert-base-uncased")
6
+
7
+ # Save model weights and tokenizer
8
+ model.save_pretrained("./my_model")
9
+ tokenizer.save_pretrained("./my_model")
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "pad_token": "[PAD]",
4
+ "sep_token": "[SEP]",
5
+ "unk_token": "[UNK]"
6
+ }
tokenizer.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "do_lower_case": true,
3
+ "max_len": 512
4
+ }