kshitijthakkar commited on
Commit
f3ea6cc
·
verified ·
1 Parent(s): d6d4945

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -82,7 +82,7 @@ datasets:
82
  ```python
83
  from transformers import AutoModelForCausalLM, AutoTokenizer
84
 
85
- tokenizer = AutoTokenizer.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b4-3060")
86
 
87
  messages = [
88
  {
@@ -98,7 +98,7 @@ messages = [
98
 
99
  tokenizer.pad_token = tokenizer.eos_token
100
  inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
101
- model = AutoModelForCausalLM.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b4-3060", device_map="auto")
102
  memory = model.get_memory_footprint() / 1e6
103
  print(f"Memory footprint: {memory:,.1f} MB")
104
  model
 
82
  ```python
83
  from transformers import AutoModelForCausalLM, AutoTokenizer
84
 
85
+ tokenizer = AutoTokenizer.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b16-3060-v2-finetuned")
86
 
87
  messages = [
88
  {
 
98
 
99
  tokenizer.pad_token = tokenizer.eos_token
100
  inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to("cuda")
101
+ model = AutoModelForCausalLM.from_pretrained("kshitijthakkar/loggenix-moe-0.12B-A0.08B-e5-lr5e4-b16-3060-v2-finetuned", device_map="auto")
102
  memory = model.get_memory_footprint() / 1e6
103
  print(f"Memory footprint: {memory:,.1f} MB")
104
  model