
sparse-encoder-testing/splade-bert-tiny-nq-onnx
Feature Extraction
•
0.0B
•
Updated
•
51
None defined yet.
backend="onnx"
or backend="openvino"
when initializing a SparseEncoder to get started, but I also included utility functions for optimization, dynamic quantization, and static quantization, plus benchmarks.n-tuple-scores
output format from mine_hard_negatives
gather_across_devices=True
to load in-batch negatives from the other devices too! Essentially a free lunch, pretty big impact potential in my evals.transformers
, and you install trackio
with pip install trackio
, then your experiments will also automatically be tracked locally with trackio. Just open up localhost and have a look at your losses/evals, no logins, no metric uploading.CrossEncoder("cross-encoder/ms-marco-MiniLM-L6-v2", backend="onnx")
export_optimized_onnx_model
, export_dynamic_quantized_onnx_model
, and export_static_quantized_openvino_model
functions now work with CrossEncoder rerankers, allowing you to optimize (e.g. fusions, gelu approximations, etc.) or quantize (int8 weights) rerankers.mine_hard_negatives
. absolute_margin
ensures that sim(query, negative) < sim(query, positive) - absolute_margin
, i.e. an absolute margin between the negative & positive similarities.relative_margin
ensures that sim(query, negative) < sim(query, positive) * (1 - relative_margin)
, i.e. a relative margin between the negative & positive similarities.export_static_quantized_openvino_model
method to quantize a model.prompts
argument in SentenceTransformerTrainingArguments
. Our experiments show that you can easily reach 0.66% to 0.90% relative performance improvement on NDCG@10 at no extra cost by adding "query: " before each training query and "document: " before each training answer.SentenceTransformer("all-MiniLM-L6-v2", backend="onnx")
. Does your model not have an ONNX or OpenVINO file yet? No worries - it'll be autoexported for you. Thank me later 😉from_model2vec
or with from_distillation
where you do the distillation yourself. It'll only take 5 seconds on GPU & 2 minutes on CPU, no dataset needed.mine_hard_negatives
docs: https://sbert.net/docs/package_reference/util.html#sentence_transformers.util.mine_hard_negatives