Update README.md
Browse files
README.md
CHANGED
@@ -82,7 +82,7 @@ datasets:
|
|
82 |
```python
|
83 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
84 |
|
85 |
-
tokenizer = AutoTokenizer.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-
|
86 |
|
87 |
messages = [
|
88 |
{
|
@@ -98,7 +98,7 @@ messages = [
|
|
98 |
|
99 |
tokenizer.pad_token = tokenizer.eos_token
|
100 |
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
|
101 |
-
model = AutoModelForCausalLM.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-
|
102 |
memory = model.get_memory_footprint() / 1e6
|
103 |
print(f"Memory footprint: {memory:,.1f} MB")
|
104 |
model
|
|
|
82 |
```python
|
83 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
84 |
|
85 |
+
tokenizer = AutoTokenizer.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b16-3060-v2-finetuned")
|
86 |
|
87 |
messages = [
|
88 |
{
|
|
|
98 |
|
99 |
tokenizer.pad_token = tokenizer.eos_token
|
100 |
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
|
101 |
+
model = AutoModelForCausalLM.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b16-3060-v2-finetuned", device_map="auto")
|
102 |
memory = model.get_memory_footprint() / 1e6
|
103 |
print(f"Memory footprint: {memory:,.1f} MB")
|
104 |
model
|