Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
2 |
license: gemma
|
3 |
base_model:
|
4 |
-
- google/embeddinggemma-
|
5 |
pipeline_tag: sentence-similarity
|
6 |
library_name: transformers.js
|
7 |
---
|
@@ -50,7 +50,7 @@ These model weights are designed to be used with [Transformers.js](https://huggi
|
|
50 |
import { AutoModel, AutoTokenizer, matmul } from "@huggingface/transformers";
|
51 |
|
52 |
// Download from the 🤗 Hub
|
53 |
-
const model_id = "onnx-community/embeddinggemma-
|
54 |
const tokenizer = await AutoTokenizer.from_pretrained(model_id);
|
55 |
const model = await AutoModel.from_pretrained(model_id, {
|
56 |
dtype: "fp32", // Options: "fp32" | "q8" | "q4".
|
@@ -97,7 +97,7 @@ import onnxruntime as ort
|
|
97 |
from transformers import AutoTokenizer
|
98 |
|
99 |
# Download from the 🤗 Hub
|
100 |
-
model_id = "onnx-community/embeddinggemma-
|
101 |
model_path = hf_hub_download(model_id, subfolder="onnx", filename="model.onnx") # Download graph
|
102 |
hf_hub_download(model_id, subfolder="onnx", filename="model.onnx_data") # Download weights
|
103 |
session = ort.InferenceSession(model_path)
|
|
|
1 |
---
|
2 |
license: gemma
|
3 |
base_model:
|
4 |
+
- google/embeddinggemma-300m
|
5 |
pipeline_tag: sentence-similarity
|
6 |
library_name: transformers.js
|
7 |
---
|
|
|
50 |
import { AutoModel, AutoTokenizer, matmul } from "@huggingface/transformers";
|
51 |
|
52 |
// Download from the 🤗 Hub
|
53 |
+
const model_id = "onnx-community/embeddinggemma-300m-ONNX";
|
54 |
const tokenizer = await AutoTokenizer.from_pretrained(model_id);
|
55 |
const model = await AutoModel.from_pretrained(model_id, {
|
56 |
dtype: "fp32", // Options: "fp32" | "q8" | "q4".
|
|
|
97 |
from transformers import AutoTokenizer
|
98 |
|
99 |
# Download from the 🤗 Hub
|
100 |
+
model_id = "onnx-community/embeddinggemma-300m-ONNX"
|
101 |
model_path = hf_hub_download(model_id, subfolder="onnx", filename="model.onnx") # Download graph
|
102 |
hf_hub_download(model_id, subfolder="onnx", filename="model.onnx_data") # Download weights
|
103 |
session = ort.InferenceSession(model_path)
|