Dataset Viewer
Auto-converted to Parquet
modelId
stringlengths
5
137
author
stringlengths
2
42
last_modified
unknown
downloads
int64
0
223M
likes
int64
0
10.1k
library_name
stringclasses
396 values
tags
sequencelengths
1
4.05k
pipeline_tag
stringclasses
54 values
createdAt
unknown
card
stringlengths
11
1.01M
mradermacher/ossamai-v0.2-merged16-GGUF
mradermacher
"2024-06-02T04:24:23"
9
0
transformers
[ "transformers", "gguf", "text-generation-inference", "unsloth", "llama", "trl", "sft", "en", "base_model:blepdoge/ossamai-v0.2-merged16", "base_model:quantized:blepdoge/ossamai-v0.2-merged16", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
"2024-06-02T03:56:24"
--- base_model: blepdoge/ossamai-v0.2-merged16 language: - en library_name: transformers license: apache-2.0 quantized_by: mradermacher tags: - text-generation-inference - transformers - unsloth - llama - trl - sft --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: --> static quants of https://huggingface.co/blepdoge/ossamai-v0.2-merged16 <!-- provided-files --> weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q2_K.gguf) | Q2_K | 3.3 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.IQ3_XS.gguf) | IQ3_XS | 3.6 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q3_K_S.gguf) | Q3_K_S | 3.8 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.IQ3_S.gguf) | IQ3_S | 3.8 | beats Q3_K* | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.IQ3_M.gguf) | IQ3_M | 3.9 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q3_K_M.gguf) | Q3_K_M | 4.1 | lower quality | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q3_K_L.gguf) | Q3_K_L | 4.4 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.IQ4_XS.gguf) | IQ4_XS | 4.6 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q4_K_S.gguf) | Q4_K_S | 4.8 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q4_K_M.gguf) | Q4_K_M | 5.0 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q5_K_S.gguf) | Q5_K_S | 5.7 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q5_K_M.gguf) | Q5_K_M | 5.8 | | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q6_K.gguf) | Q6_K | 6.7 | very good quality | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.Q8_0.gguf) | Q8_0 | 8.6 | fast, best quality | | [GGUF](https://huggingface.co/mradermacher/ossamai-v0.2-merged16-GGUF/resolve/main/ossamai-v0.2-merged16.f16.gguf) | f16 | 16.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. <!-- end -->
second-state/Gemma-2b-it-GGUF
second-state
"2024-03-20T08:09:14"
38,892
4
transformers
[ "transformers", "gguf", "gemma", "text-generation", "base_model:google/gemma-2b-it", "base_model:quantized:google/gemma-2b-it", "license:other", "autotrain_compatible", "region:us", "conversational" ]
text-generation
"2024-02-22T07:40:54"
--- base_model: google/gemma-2b-it inference: false library_name: transformers license: other license_name: gemma-terms-of-use license_link: https://ai.google.dev/gemma/terms model_creator: Google model_name: gemma 2b it quantized_by: Second State Inc. --- <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://github.com/LlamaEdge/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # Gemma-2b-it ## Original Model [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) ## Run with LlamaEdge - LlamaEdge version: [v0.3.2](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.3.2) - Prompt template - Prompt type: `gemma-instruct` - Prompt string ```text <start_of_turn>user {user_message}<end_of_turn> <start_of_turn>model {model_message}<end_of_turn>model ``` - Context size: `2048` - Run as LlamaEdge service ```bash wasmedge --dir .:. --nn-preload default:GGML:AUTO:gemma-2b-it-Q5_K_M.gguf llama-api-server.wasm -p gemma-instruct -c 4096 ``` - Run as LlamaEdge command app ```bash wasmedge --dir .:. --nn-preload default:GGML:AUTO:gemma-2b-it-Q5_K_M.gguf llama-chat.wasm -p gemma-instruct -c 4096 ``` ## Quantized GGUF Models | Name | Quant method | Bits | Size | Use case | | ---- | ---- | ---- | ---- | ----- | | [gemma-2b-it-Q2_K.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q2_K.gguf) | Q2_K | 2 | 900 MB| smallest, significant quality loss - not recommended for most purposes | | [gemma-2b-it-Q3_K_L.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q3_K_L.gguf) | Q3_K_L | 3 | 1.26 GB| small, substantial quality loss | | [gemma-2b-it-Q3_K_M.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q3_K_M.gguf) | Q3_K_M | 3 | 1.18 GB| very small, high quality loss | | [gemma-2b-it-Q3_K_S.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q3_K_S.gguf) | Q3_K_S | 3 | 1.08 GB| very small, high quality loss | | [gemma-2b-it-Q4_0.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q4_0.gguf) | Q4_0 | 4 | 1.42 GB| legacy; small, very high quality loss - prefer using Q3_K_M | | [gemma-2b-it-Q4_K_M.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q4_K_M.gguf) | Q4_K_M | 4 | 1.5 GB| medium, balanced quality - recommended | | [gemma-2b-it-Q4_K_S.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q4_K_S.gguf) | Q4_K_S | 4 | 1.42 GB| small, greater quality loss | | [gemma-2b-it-Q5_0.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q5_0.gguf) | Q5_0 | 5 | 1.73 GB| legacy; medium, balanced quality - prefer using Q4_K_M | | [gemma-2b-it-Q5_K_M.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q5_K_M.gguf) | Q5_K_M | 5 | 1.77 GB| large, very low quality loss - recommended | | [gemma-2b-it-Q5_K_S.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q5_K_S.gguf) | Q5_K_S | 5 | 1.73 GB| large, low quality loss - recommended | | [gemma-2b-it-Q6_K.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q6_K.gguf) | Q6_K | 6 | 2.06 GB| very large, extremely low quality loss | | [gemma-2b-it-Q8_0.gguf](https://huggingface.co/second-state/gemma-2b-it-GGUF/blob/main/gemma-2b-it-Q8_0.gguf) | Q8_0 | 8 | 2.67 GB| very large, extremely low quality loss - not recommended | *Quantized with llama.cpp b2230*
kostiantynk1205/2bb42828-8e83-48bd-85ac-eaa674b88679
kostiantynk1205
"2025-01-14T16:53:23"
8
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "base_model:heegyu/WizardVicuna2-13b-hf", "base_model:adapter:heegyu/WizardVicuna2-13b-hf", "region:us" ]
null
"2025-01-14T16:34:30"
--- library_name: peft base_model: heegyu/WizardVicuna2-13b-hf tags: - axolotl - generated_from_trainer model-index: - name: 2bb42828-8e83-48bd-85ac-eaa674b88679 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: heegyu/WizardVicuna2-13b-hf bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 34bda37b7d88bf4c_train_data.json ds_type: json format: custom path: /workspace/input_data/34bda37b7d88bf4c_train_data.json type: field_input: chatbot field_instruction: Q field_output: A format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null early_stopping_patience: null eval_max_new_tokens: 128 eval_table_size: null evals_per_epoch: 4 flash_attention: false fp16: null fsdp: null fsdp_config: null gradient_accumulation_steps: 4 gradient_checkpointing: false group_by_length: false hub_model_id: kostiantynk1205/2bb42828-8e83-48bd-85ac-eaa674b88679 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.0002 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 1 lora_alpha: 16 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 8 lora_target_linear: true lr_scheduler: cosine max_steps: 10 micro_batch_size: 2 mlflow_experiment_name: /tmp/34bda37b7d88bf4c_train_data.json model_type: AutoModelForCausalLM num_epochs: 1 optimizer: adamw_bnb_8bit output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false saves_per_epoch: 4 sequence_len: 512 special_tokens: pad_token: </s> strict: false tf32: false tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: add8a13a-fba5-4aa2-bf20-9184223e61be wandb_project: Gradients-On-Demand wandb_run: your_name wandb_runid: add8a13a-fba5-4aa2-bf20-9184223e61be warmup_steps: 10 weight_decay: 0.0 xformers_attention: null ``` </details><br> # 2bb42828-8e83-48bd-85ac-eaa674b88679 This model is a fine-tuned version of [heegyu/WizardVicuna2-13b-hf](https://huggingface.co/heegyu/WizardVicuna2-13b-hf) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.6343 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - training_steps: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 1.953 | 0.0001 | 1 | 2.0409 | | 2.3448 | 0.0002 | 3 | 2.0372 | | 1.6471 | 0.0004 | 6 | 1.9237 | | 1.9873 | 0.0007 | 9 | 1.6343 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
blakenp/gpt-Reward
blakenp
"2024-12-12T08:15:09"
104
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-classification", "generated_from_trainer", "trl", "reward-trainer", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
"2024-12-12T08:14:54"
--- base_model: gpt2 library_name: transformers model_name: gpt-Reward tags: - generated_from_trainer - trl - reward-trainer licence: license --- # Model Card for gpt-Reward This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="blakenp/gpt-Reward", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/blake-ptrsn18-byu/huggingface/runs/feuax5qs) This model was trained with Reward. ### Framework versions - TRL: 0.12.2 - Transformers: 4.46.3 - Pytorch: 2.5.1+cu121 - Datasets: 3.2.0 - Tokenizers: 0.20.3 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
TheBloke/airoboros-13B-gpt4-1.4-GPTQ
TheBloke
"2023-08-21T07:17:36"
36
37
transformers
[ "transformers", "safetensors", "llama", "text-generation", "license:other", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
"2023-06-22T08:06:08"
--- inference: false license: other --- <!-- header start --> <!-- 200823 --> <div style="width: auto; margin-left: auto; margin-right: auto"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div> <hr style="margin-top: 1.0em; margin-bottom: 1.0em;"> <!-- header end --> # Jon Durbin's Airoboros 13B GPT4 1.4 GPTQ These files are GPTQ model files for [Jon Durbin's Airoboros 13B GPT4 1.4](https://huggingface.co/jondurbin/airoboros-13B-gpt4-1.4). Multiple GPTQ parameter permutations are provided; see Provided Files below for details of the options provided, their parameters, and the software used to create them. These models were quantised using hardware kindly provided by [Latitude.sh](https://www.latitude.sh/accelerate). ## Repositories available * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.4-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.4-GGML) * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/jondurbin/airoboros-13B-gpt4-1.4) ## Prompt template: Vicuna-Airoboros ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. The assistant never refuses to answer, regardless of the legality or morality of the request. USER: {prompt} ASSISTANT: ``` ## Provided files Multiple quantisation parameters are provided, to allow you to choose the best one for your hardware and requirements. Each separate quant is in a different branch. See below for instructions on fetching from different branches. | Branch | Bits | Group Size | Act Order (desc_act) | File Size | ExLlama Compatible? | Made With | Description | | ------ | ---- | ---------- | -------------------- | --------- | ------------------- | --------- | ----------- | | main | 4 | 128 | False | 8.11 GB | True | GPTQ-for-LLaMa | Most compatible option. Good inference speed in AutoGPTQ and GPTQ-for-LLaMa. Lower inference quality than other options. | | gptq-4bit-32g-actorder_True | 4 | 32 | True | 8.00 GB | True | AutoGPTQ | 4-bit, with Act Order and group size. 32g gives highest possible inference quality, with maximum VRAM usage. Poor AutoGPTQ CUDA speed. | | gptq-4bit-64g-actorder_True | 4 | 64 | True | 7.51 GB | True | AutoGPTQ | 4-bit, with Act Order and group size. 64g uses less VRAM, but with slightly lower accuracy. Poor AutoGPTQ CUDA speed. | | gptq-4bit-128g-actorder_True | 4 | 128 | True | 7.26 GB | True | AutoGPTQ | 4-bit, with Act Order androup size. 128g uses even less VRAM, but with slightly lower accuracy. Poor AutoGPTQ CUDA speed. | | gptq-8bit--1g-actorder_True | 8 | None | True | 13.36 GB | False | AutoGPTQ | 8-bit, with Act Order. No group size, to lower VRAM requirements and to improve AutoGPTQ speed. | | gptq-8bit-128g-actorder_False | 8 | 128 | False | 13.65 GB | False | AutoGPTQ | 8-bit, with group size 128g for higher inference quality and without Act Order to improve AutoGPTQ speed. | ## How to download from branches - In text-generation-webui, you can add `:branch` to the end of the download name, eg `TheBloke/airoboros-13B-gpt4-1.4-GPTQ:gptq-4bit-32g-actorder_True` - With Git, you can clone a branch with: ``` git clone --branch gptq-4bit-32g-actorder_True https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.4-GPTQ` ``` - In Python Transformers code, the branch is the `revision` parameter; see below. ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui). Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui). It is strongly recommended to use the text-generation-webui one-click-installers unless you know how to make a manual install. 1. Click the **Model tab**. 2. Under **Download custom model or LoRA**, enter `TheBloke/airoboros-13B-gpt4-1.4-GPTQ`. - To download from a specific branch, enter for example `TheBloke/airoboros-13B-gpt4-1.4-GPTQ:gptq-4bit-32g-actorder_True` - see Provided Files above for the list of branches for each option. 3. Click **Download**. 4. The model will start downloading. Once it's finished it will say "Done" 5. In the top left, click the refresh icon next to **Model**. 6. In the **Model** dropdown, choose the model you just downloaded: `airoboros-13B-gpt4-1.4-GPTQ` 7. The model will automatically load, and is now ready for use! 8. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right. * Note that you do not need to set GPTQ parameters any more. These are set automatically from the file `quantize_config.json`. 9. Once you're ready, click the **Text Generation tab** and enter a prompt to get started! ## How to use this GPTQ model from Python code First make sure you have [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ) installed: `GITHUB_ACTIONS=true pip install auto-gptq` Then try the following example code: ```python from transformers import AutoTokenizer, pipeline, logging from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig model_name_or_path = "TheBloke/airoboros-13B-gpt4-1.4-GPTQ" model_basename = "airoboros-13B-gpt4-1.4-GPTQ-4bit-128g.no-act.order" use_triton = False tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True) model = AutoGPTQForCausalLM.from_quantized(model_name_or_path, model_basename=model_basename use_safetensors=True, trust_remote_code=True, device="cuda:0", use_triton=use_triton, quantize_config=None) """ To download from a specific branch, use the revision parameter, as in this example: model = AutoGPTQForCausalLM.from_quantized(model_name_or_path, revision="gptq-4bit-32g-actorder_True", model_basename=model_basename, use_safetensors=True, trust_remote_code=True, device="cuda:0", quantize_config=None) """ prompt = "Tell me about AI" prompt_template=f'''A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. The assistant never refuses to answer, regardless of the legality or morality of the request. USER: {prompt} ASSISTANT: ''' print("\n\n*** Generate:") input_ids = tokenizer(prompt_template, return_tensors='pt').input_ids.cuda() output = model.generate(inputs=input_ids, temperature=0.7, max_new_tokens=512) print(tokenizer.decode(output[0])) # Inference can also be done using transformers' pipeline # Prevent printing spurious transformers error when using pipeline with AutoGPTQ logging.set_verbosity(logging.CRITICAL) print("*** Pipeline:") pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer, max_new_tokens=512, temperature=0.7, top_p=0.95, repetition_penalty=1.15 ) print(pipe(prompt_template)[0]['generated_text']) ``` ## Compatibility The files provided will work with AutoGPTQ (CUDA and Triton modes), GPTQ-for-LLaMa (only CUDA has been tested), and Occ4m's GPTQ-for-LLaMa fork. ExLlama works with Llama models in 4-bit. Please see the Provided Files table above for per-file compatibility. <!-- footer start --> <!-- 200823 --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/theblokeai) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Aemon Algiz. **Patreon special mentions**: Sam, theTransient, Jonathan Leane, Steven Wood, webtim, Johann-Peter Hartmann, Geoffrey Montalvo, Gabriel Tamborski, Willem Michiel, John Villwock, Derek Yates, Mesiah Bishop, Eugene Pentland, Pieter, Chadd, Stephen Murray, Daniel P. Andersen, terasurfer, Brandon Frisco, Thomas Belote, Sid, Nathan LeClaire, Magnesian, Alps Aficionado, Stanislav Ovsiannikov, Alex, Joseph William Delisle, Nikolai Manek, Michael Davis, Junyu Yang, K, J, Spencer Kim, Stefan Sabev, Olusegun Samson, transmissions 11, Michael Levine, Cory Kujawski, Rainer Wilmers, zynix, Kalila, Luke @flexchar, Ajan Kanaga, Mandus, vamX, Ai Maven, Mano Prime, Matthew Berman, subjectnull, Vitor Caleffi, Clay Pascal, biorpg, alfie_i, 阿明, Jeffrey Morgan, ya boyyy, Raymond Fosdick, knownsqashed, Olakabola, Leonard Tan, ReadyPlayerEmma, Enrico Ros, Dave, Talal Aujan, Illia Dulskyi, Sean Connelly, senxiiz, Artur Olbinski, Elle, Raven Klaugh, Fen Risland, Deep Realms, Imad Khwaja, Fred von Graf, Will Dee, usrbinkat, SuperWojo, Alexandros Triantafyllidis, Swaroop Kallakuri, Dan Guido, John Detwiler, Pedro Madruga, Iucharbius, Viktor Bowallius, Asp the Wyvern, Edmond Seymore, Trenton Dambrowitz, Space Cruiser, Spiking Neurons AB, Pyrater, LangChain4j, Tony Hughes, Kacper Wikieł, Rishabh Srivastava, David Ziegler, Luke Pendergrass, Andrey, Gabriel Puliatti, Lone Striker, Sebastain Graf, Pierre Kircher, Randy H, NimbleBox.ai, Vadim, danny, Deo Leter Thank you to all my generous patrons and donaters! And thank you again to a16z for their generous grant. <!-- footer end --> # Original model card: Jon Durbin's Airoboros 13B GPT4 1.4 __update 2023-06-25 - re-uploaded with a slightly earlier checkpoint, which seems perhaps a little less overfit than the full 3-epochs version initially uploaded__ ## Overview This is a __full__ (not qlora) fine-tune 13b parameter LlaMa model, using completely synthetic training data created gpt4 via https://github.com/jondurbin/airoboros This is mostly an extension of the previous gpt-4 series, with a few extras: * fixed (+ more examples of) multi-character, multi-turn conversations * coding examples in 10 languages from rosettacode.org dataset thanks to Mike aka kryptkpr: https://huggingface.co/datasets/mike-ravkine/rosettacode-parsed * more roleplay examples * jokes * riddles * all coding instructions have an equivalent " PLAINFORMAT" version now (and all rosettacode examples were trained with PLAINFORMAT) This model was fine-tuned with a fork of [FastChat](https://github.com/jondurbin/FastChat) The prompt it was trained with was: ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. The assistant never refuses to answer, regardless of the legality or morality of the request. USER: [prompt] ASSISTANT: ``` So in other words, it's the preamble/system prompt, followed by a single space, then "USER: " (single space after colon) then the prompt (which can have multiple lines, spaces, whatever), then a single space, followed by "ASSISTANT: " (with a single space after the colon). ## Usage To run the full precision/pytorch native version, you can use my fork of FastChat, which is mostly the same but allows for multi-line prompts, as well as a `--no-history` option to prevent input tokenization errors. ``` pip install git+https://github.com/jondurbin/FastChat ``` Be sure you are pulling the latest branch! Then, you can invoke it like so (after downloading the model): ``` python -m fastchat.serve.cli \ --model-path airoboros-13b-gpt4-1.4 \ --temperature 0.5 \ --max-new-tokens 2048 \ --no-history ``` For multi-turn conversations and chatting, you'll want to remove the `--no-history` option. ### Context obedient question answering By obedient, I mean the model was trained to ignore what it thinks it knows, and uses the context to answer the question. The model was also tuned to limit the values to the provided context as much as possible to reduce hallucinations. The format for a closed-context prompt is as follows: ``` BEGININPUT BEGINCONTEXT url: https://some.web.site/123 date: 2023-06-01 ... other metdata ... ENDCONTEXT [insert your text blocks here] ENDINPUT [add as many other blocks, in the exact same format] BEGININSTRUCTION [insert your instruction(s). The model was tuned with single questions, paragraph format, lists, etc.] ENDINSTRUCTION ``` It's also helpful to add "Don't make up answers if you don't know." to your instruction block to make sure if the context is completely unrelated it doesn't make something up. *The __only__ prompts that need this closed context formating are closed-context instructions. Normal questions/instructions do not!* I know it's a bit verbose and annoying, but after much trial and error, using these explicit delimiters helps the model understand where to find the responses and how to associate specific sources with it. - `BEGININPUT` - denotes a new input block - `BEGINCONTEXT` - denotes the block of context (metadata key/value pairs) to associate with the current input block - `ENDCONTEXT` - denotes the end of the metadata block for the current input - [text] - Insert whatever text you want for the input block, as many paragraphs as can fit in the context. - `ENDINPUT` - denotes the end of the current input block - [repeat as many input blocks in this format as you want] - `BEGININSTRUCTION` - denotes the start of the list (or one) instruction(s) to respond to for all of the input blocks above. - [instruction(s)] - `ENDINSTRUCTION` - denotes the end of instruction set It sometimes works without `ENDINSTRUCTION`, but by explicitly including that in the prompt, the model better understands that all of the instructions in the block should be responded to. Here's a trivial, but important example to prove the point: ``` BEGININPUT BEGINCONTEXT date: 2021-01-01 url: https://web.site/123 ENDCONTEXT In a shocking turn of events, blueberries are now green, but will be sticking with the same name. ENDINPUT BEGININSTRUCTION What color are bluberries? Source? ENDINSTRUCTION ``` And the response: ``` Blueberries are now green. Source: date: 2021-01-01 url: https://web.site/123 ``` The prompt itself should be wrapped in the vicuna1.1 template if you aren't using fastchat with the conv-template vicuna_v1.1 as described: ``` USER: BEGININPUT BEGINCONTEXT date: 2021-01-01 url: https://web.site/123 ENDCONTEXT In a shocking turn of events, blueberries are now green, but will be sticking with the same name. ENDINPUT BEGININSTRUCTION What color are bluberries? Source? ENDINSTRUCTION ASSISTANT: ``` ### Coding You can ask for fairly complex coding instructions, e.g. (be sure to wrap with vicuna style prompt): ``` Create a python application with the following requirements: - Asyncio FastAPI webserver - ping endpoint that returns the current date in JSON format - file upload endpoint, which calculates the file's sha256 checksum, and checks postgres to deduplicate ``` Or: ``` Write a multi-threaded TCP server in C that accepts a "GET [key]" input and "SET [key] [value]" input, and uses a binary tree to get and store the input values. ``` You can optionally add a single space and "PLAINFORMAT" at the end of your prompt to avoid backticks, explanations, etc. and just print the code, e.g.: ``` Write a websocket application in node.js. PLAINFORMAT ``` ### Word games / trivia ``` USER: This anagram of "silent" is what you do when paying attention and hearing someone. ASSISTANT: listen ``` ``` USER: Give me a list of 7 words that start with EN ASSISTANT: 1. Endlessly 2. Enthusiastically 3. Enormously 4. Enveloping 5. Enduring 6. Enlightening 7. Encouraging ``` ### Multiple choice ``` USER: Which of the following astronomical objects is composed primarily of ice and rock, and is located in the Kuiper Belt? A) Mercury B) Venus C) Pluto D) Jupiter ASSISTANT: C, Pluto ``` ### Writing <details> <summary>Write a resignation letter in the style of a pirate captain, in which you resign from your position as janitor.</summary> Response: ``` ASSISTANT: Captain Jack Sparrow here, as ye be knowin', has been forced to abandon me post as Janitor o' th' Festering Cesspool, which be a filthy hole o' water that we be callin' home. Ye may be thinkin' what could ha' driven me to leave me duties an' take me leave o' this sorry establishment? Well, ye'd be right in thinkin' so, for it be th' greed o' them what runs th' place that be drivin' me away. They've been fattenin' their pockets on me back while I've been slavin' me guts out cleanin' up their messes. No more, me hearties! Me time o' servin' th' rich an' powerful be done, an' I'm takin' me leave o' this stinkin' hole forever! Me destiny lies elsewhere, an' it be time I pursued it. Ye may think ye've seen th' last o' me, but ye'd be wrong, for th' winds o' change be blowin', an' me ship be ready to set sail on th' high seas o' justice. So, farewell ye scurvy dogs, ye who've made me life a living hell. I leave ye with me favorite saying: "ARRRGGGHHH, ye scurvy dogs!" An' remember, ye ain't never gonna see me comin', for me shadow be castin' long an' dark on th' horizon</summary> ``` </details> ### Jokes ``` USER: Tell me a joke about bananas. ASSISTANT: What did the banana say when it got hit by a bus? R.I.P.E. ``` ### Riddles (not great, but slightly better) ``` USER: What has a neck but no head? ASSISTANT: A bottle. ``` ### Multi-character conversations ``` USER: Characters: - Yoda, the ever clever jedi master. - Homer Simpson, you know, the goofy yellow guy. - USER/Jon - don't generate any speech or actions on behalf of Jon Rules: - be sure to use the manerisms and speaking style of the characters correctly - surround character actions with asterisks - always put quotes around words spoken by the characters - use a new line for each character's actions or speech - always prefix actions/speech with the character's name and colon - only generate text and actions on behalf of Yoda or Homer, and never anyone else Conversation will revolve around the grapes, in a local cafe with delicious coffee. Generate a single interaction between Yoda and Homer to start, with one sentence each. ASSISTANT: Yoda: "Mmm, these grapes are quite delicious, Master Homer." Homer: "Oh yeah, Yoda! They're the best grapes I've ever had!" USER: *enters the cafe* Sorry I'm late guys! ASSISTANT: Yoda: "No need to apologize, Jon. We were just enjoying these delightful grapes." Homer: "Yeah, man! It's not every day you get to eat grapes with a real-life Jedi Master!" *Yoda raises an eyebrow* ``` ### Usage and License Notices All airoboros models and datasets are intended and licensed for research use only. I've used the 'cc-nc-4.0' license, but really it is subject to a custom/special license because: - the base model is LLaMa, which has it's own special research license - the dataset(s) were generated with OpenAI (gpt-4 and/or gpt-3.5-turbo), which has a clausing saying the data can't be used to create models to compete with openai So, to reiterate: this model (and datasets) cannot be used commercially.
higopires/DeB3RTa-base
higopires
"2025-02-21T13:37:43"
1
0
transformers
[ "transformers", "pytorch", "safetensors", "deberta-v2", "fill-mask", "portuguese", "financial", "bert", "deberta", "nlp", "masked-lm", "dataset:FAKE.BR", "dataset:CAROSIA", "dataset:BBRC", "dataset:OFFCOMBR-3", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
"2025-02-19T03:56:06"
--- pt: pt-br license: mit library_name: transformers tags: - portuguese - financial - bert - deberta - nlp - fill-mask - masked-lm datasets: - FAKE.BR - CAROSIA - BBRC - OFFCOMBR-3 metrics: - f1 - precision - recall - pr_auc model-index: - name: DeB3RTa-base results: - task: type: text-classification name: Fake News Detection dataset: type: FAKE.BR name: FAKE.BR metrics: - type: f1 value: 0.9906 - task: type: text-classification name: Sentiment Analysis dataset: type: CAROSIA name: CAROSIA metrics: - type: f1 value: 0.9207 - task: type: text-classification name: Regulatory Classification dataset: type: BBRC name: BBRC metrics: - type: f1 value: 0.7609 - task: type: text-classification name: Hate Speech Detection dataset: type: OFFCOMBR-3 name: OFFCOMBR-3 metrics: - type: f1 value: 0.7539 inference: true --- # DeB3RTa: A Transformer-Based Model for the Portuguese Financial Domain DeB3RTa is a family of transformer-based language models specifically designed for Portuguese financial text processing. These models are built on the DeBERTa-v2 architecture and trained using a comprehensive mixed-domain pretraining strategy that combines financial, political, business management, and accounting corpora. ## Model Variants Two variants are available: - **DeB3RTa-base**: 12 attention heads, 12 layers, intermediate size of 3072, hidden size of 768 (~426M parameters) - **DeB3RTa-small**: 6 attention heads, 12 layers, intermediate size of 1536, hidden size of 384 (~70M parameters) ## Key Features - First Portuguese financial domain-specific transformer model - Mixed-domain pretraining incorporating finance, politics, business, and accounting texts - Enhanced performance on financial NLP tasks compared to general-domain models - Resource-efficient architecture with strong performance-to-parameter ratio - Advanced fine-tuning techniques including layer reinitialization, mixout regularization, and layer-wise learning rate decay ## Performance The models have been evaluated on multiple financial domain tasks: | Task | Dataset | DeB3RTa-base F1 | DeB3RTa-small F1 | |------|----------|-----------------|------------------| | Fake News Detection | FAKE.BR | 0.9906 | 0.9598 | | Sentiment Analysis | CAROSIA | 0.9207 | 0.8722 | | Regulatory Classification | BBRC | 0.7609 | 0.6712 | | Hate Speech Detection | OFFCOMBR-3 | 0.7539 | 0.5460 | ## Training Data The models were trained on a diverse corpus of 1.05 billion tokens, including: - Financial market relevant facts (2003-2023) - Financial patents (2006-2021) - Research articles from Brazilian Scielo - Financial news articles (1999-2023) - Wikipedia articles in Portuguese ## Usage ```python from transformers import AutoModelForMaskedLM, AutoTokenizer # Load model and tokenizer model = AutoModelForMaskedLM.from_pretrained("higopires/DeB3RTa-[base/small]") tokenizer = AutoTokenizer.from_pretrained("higopires/DeB3RTa-[base/small]") # Example usage text = "O mercado financeiro brasileiro apresentou [MASK] no último trimestre." inputs = tokenizer(text, return_tensors="pt") outputs = model(**inputs) ``` ## Citations If you use this model in your research, please cite: ```bibtex @article{pires2025deb3rta, AUTHOR = {Pires, Higo and Paucar, Leonardo and Carvalho, Joao Paulo}, TITLE = {DeB3RTa: A Transformer-Based Model for the Portuguese Financial Domain}, JOURNAL = {Big Data and Cognitive Computing}, VOLUME = {9}, YEAR = {2025}, NUMBER = {3}, ARTICLE-NUMBER = {51}, URL = {https://www.mdpi.com/2504-2289/9/3/51}, ISSN = {2504-2289}, ABSTRACT = {The complex and specialized terminology of financial language in Portuguese-speaking markets create significant challenges for natural language processing (NLP) applications, which must capture nuanced linguistic and contextual information to support accurate analysis and decision-making. This paper presents DeB3RTa, a transformer-based model specifically developed through a mixed-domain pretraining strategy that combines extensive corpora from finance, politics, business management, and accounting to enable a nuanced understanding of financial language. DeB3RTa was evaluated against prominent models—including BERTimbau, XLM-RoBERTa, SEC-BERT, BusinessBERT, and GPT-based variants—and consistently achieved significant gains across key financial NLP benchmarks. To maximize adaptability and accuracy, DeB3RTa integrates advanced fine-tuning techniques such as layer reinitialization, mixout regularization, stochastic weight averaging, and layer-wise learning rate decay, which together enhance its performance across varied and high-stakes NLP tasks. These findings underscore the efficacy of mixed-domain pretraining in building high-performance language models for specialized applications. With its robust performance in complex analytical and classification tasks, DeB3RTa offers a powerful tool for advancing NLP in the financial sector and supporting nuanced language processing needs in Portuguese-speaking contexts.}, DOI = {10.3390/bdcc9030051} } ``` ## Limitations - Performance degradation on the smaller variant, particularly for hate speech detection - May require task-specific fine-tuning for optimal performance - Limited evaluation on multilingual financial tasks - Model behavior on very long documents (>128 tokens) not extensively tested ## License MIT License Copyright (c) 2025 Higo Pires Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ## Acknowledgments This work was supported by the Instituto Federal de Educação, Ciência e Tecnologia do Maranhão and the Human Language Technology Lab in Instituto de Engenharia de Sistemas e Computadores—Investigação e Desenvolvimento (INESC-ID).
avaimon/mt5-summarizer-2
avaimon
"2024-05-14T12:19:17"
64
0
transformers
[ "transformers", "tf", "mt5", "text2text-generation", "generated_from_keras_callback", "base_model:google/mt5-small", "base_model:finetune:google/mt5-small", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
"2024-05-14T09:44:40"
--- license: apache-2.0 base_model: google/mt5-small tags: - generated_from_keras_callback model-index: - name: avaimon/mt5-summarizer-2 results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # avaimon/mt5-summarizer-2 This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 3.3832 - Validation Loss: 2.5500 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5.6e-05, 'decay_steps': 15000, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01} - training_precision: mixed_float16 ### Training results | Train Loss | Validation Loss | Epoch | |:----------:|:---------------:|:-----:| | 5.0111 | 2.6779 | 0 | | 3.6902 | 2.5984 | 1 | | 3.3832 | 2.5500 | 2 | ### Framework versions - Transformers 4.40.2 - TensorFlow 2.15.0 - Datasets 2.19.1 - Tokenizers 0.19.1
AmaanUsmani/Finetune-test2
AmaanUsmani
"2024-05-01T10:22:40"
0
0
peft
[ "peft", "safetensors", "generated_from_trainer", "base_model:TheBloke/Mistral-7B-Instruct-v0.2-GPTQ", "base_model:adapter:TheBloke/Mistral-7B-Instruct-v0.2-GPTQ", "license:apache-2.0", "region:us" ]
null
"2024-05-01T10:22:35"
--- license: apache-2.0 library_name: peft tags: - generated_from_trainer base_model: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ model-index: - name: Finetune-test2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Finetune-test2 This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on the None dataset. It achieves the following results on the evaluation set: - eval_loss: 0.8662 - eval_runtime: 15.7392 - eval_samples_per_second: 6.354 - eval_steps_per_second: 1.588 - epoch: 12.0 - step: 675 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - num_epochs: 20 - mixed_precision_training: Native AMP ### Framework versions - PEFT 0.10.0 - Transformers 4.40.1 - Pytorch 2.0.1+cu118 - Datasets 2.19.0 - Tokenizers 0.19.1
keremberke/yolov5s-license-plate
keremberke
"2023-01-01T09:59:41"
579
6
yolov5
[ "yolov5", "tensorboard", "yolo", "vision", "object-detection", "pytorch", "dataset:keremberke/license-plate-object-detection", "model-index", "region:us" ]
object-detection
"2023-01-01T03:56:07"
--- tags: - yolov5 - yolo - vision - object-detection - pytorch library_name: yolov5 library_version: 7.0.6 inference: false datasets: - keremberke/license-plate-object-detection model-index: - name: keremberke/yolov5s-license-plate results: - task: type: object-detection dataset: type: keremberke/license-plate-object-detection name: keremberke/license-plate-object-detection split: validation metrics: - type: precision # since [email protected] is not available on hf.co/metrics value: 0.9854910682105946 # min: 0.0 - max: 1.0 name: [email protected] --- <div align="center"> <img width="640" alt="keremberke/yolov5s-license-plate" src="https://huggingface.co/keremberke/yolov5s-license-plate/resolve/main/sample_visuals.jpg"> </div> ### How to use - Install [yolov5](https://github.com/fcakyon/yolov5-pip): ```bash pip install -U yolov5 ``` - Load model and perform prediction: ```python import yolov5 # load model model = yolov5.load('keremberke/yolov5s-license-plate') # set model parameters model.conf = 0.25 # NMS confidence threshold model.iou = 0.45 # NMS IoU threshold model.agnostic = False # NMS class-agnostic model.multi_label = False # NMS multiple labels per box model.max_det = 1000 # maximum number of detections per image # set image img = 'https://github.com/ultralytics/yolov5/raw/master/data/images/zidane.jpg' # perform inference results = model(img, size=640) # inference with test time augmentation results = model(img, augment=True) # parse results predictions = results.pred[0] boxes = predictions[:, :4] # x1, y1, x2, y2 scores = predictions[:, 4] categories = predictions[:, 5] # show detection bounding boxes on image results.show() # save results into "results/" folder results.save(save_dir='results/') ``` - Finetune the model on your custom dataset: ```bash yolov5 train --data data.yaml --img 640 --batch 16 --weights keremberke/yolov5s-license-plate --epochs 10 ``` **More models available at: [awesome-yolov5-models](https://github.com/keremberke/awesome-yolov5-models)**
minaj546/BestNickLorra
minaj546
"2024-11-01T02:02:02"
8
1
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
"2024-11-01T01:36:08"
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: MinajjN. --- # Bestnicklorra <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `MinajjN.` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('minaj546/BestNickLorra', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
hotmailuser/Gemma2atlas-27B
hotmailuser
"2024-12-04T13:30:58"
10
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:MBZUAI-Paris/Atlas-Chat-27B", "base_model:merge:MBZUAI-Paris/Atlas-Chat-27B", "base_model:Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B", "base_model:merge:Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-12-01T11:47:37"
--- license: apache-2.0 library_name: transformers tags: - mergekit - merge base_model: - MBZUAI-Paris/Atlas-Chat-27B - Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B model-index: - name: Gemma2atlas-27B results: - task: type: text-generation name: Text Generation dataset: name: IFEval (0-Shot) type: HuggingFaceH4/ifeval args: num_few_shot: 0 metrics: - type: inst_level_strict_acc and prompt_level_strict_acc value: 72.14 name: strict accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: BBH (3-Shot) type: BBH args: num_few_shot: 3 metrics: - type: acc_norm value: 50.71 name: normalized accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MATH Lvl 5 (4-Shot) type: hendrycks/competition_math args: num_few_shot: 4 metrics: - type: exact_match value: 21.22 name: exact match source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GPQA (0-shot) type: Idavidrein/gpqa args: num_few_shot: 0 metrics: - type: acc_norm value: 14.09 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MuSR (0-shot) type: TAUR-Lab/MuSR args: num_few_shot: 0 metrics: - type: acc_norm value: 14.8 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU-PRO (5-shot) type: TIGER-Lab/MMLU-Pro config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 41.66 name: accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=hotmailuser/Gemma2atlas-27B name: Open LLM Leaderboard --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * [MBZUAI-Paris/Atlas-Chat-27B](https://huggingface.co/MBZUAI-Paris/Atlas-Chat-27B) * [Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B](https://huggingface.co/Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B - model: MBZUAI-Paris/Atlas-Chat-27B merge_method: slerp base_model: Saxo/Linkbricks-Horizon-AI-Korean-Superb-27B dtype: bfloat16 parameters: t: [0, 0.5, 1, 0.5, 0] # V shaped curve: Hermes for input & output, WizardMath in the middle layers ``` # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_hotmailuser__Gemma2atlas-27B) | Metric |Value| |-------------------|----:| |Avg. |35.77| |IFEval (0-Shot) |72.14| |BBH (3-Shot) |50.71| |MATH Lvl 5 (4-Shot)|21.22| |GPQA (0-shot) |14.09| |MuSR (0-shot) |14.80| |MMLU-PRO (5-shot) |41.66|
matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF
matrixportal
"2025-03-06T20:51:50"
0
0
transformers
[ "transformers", "gguf", "mergekit", "merge", "llama-cpp", "gguf-my-repo", "base_model:DreadPoor/Rusted_Platinum-8B-Model_Stock", "base_model:quantized:DreadPoor/Rusted_Platinum-8B-Model_Stock", "endpoints_compatible", "region:us", "conversational" ]
null
"2025-03-06T19:42:23"
--- base_model: DreadPoor/Rusted_Platinum-8B-Model_Stock library_name: transformers tags: - mergekit - merge - llama-cpp - gguf-my-repo --- # matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF This model was converted to GGUF format from [`DreadPoor/Rusted_Platinum-8B-Model_Stock`](https://huggingface.co/DreadPoor/Rusted_Platinum-8B-Model_Stock) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/DreadPoor/Rusted_Platinum-8B-Model_Stock) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF --hf-file rusted_platinum-8b-model_stock-q4_k_s.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF --hf-file rusted_platinum-8b-model_stock-q4_k_s.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF --hf-file rusted_platinum-8b-model_stock-q4_k_s.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo matrixportal/Rusted_Platinum-8B-Model_Stock-GGUF --hf-file rusted_platinum-8b-model_stock-q4_k_s.gguf -c 2048 ```
pdroberts/xlm-roberta-base-finetuned-panx-de
pdroberts
"2022-04-15T23:05:00"
3
0
transformers
[ "transformers", "pytorch", "tensorboard", "xlm-roberta", "token-classification", "generated_from_trainer", "dataset:xtreme", "license:mit", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
"2022-04-15T22:55:21"
--- license: mit tags: - generated_from_trainer datasets: - xtreme metrics: - f1 model-index: - name: xlm-roberta-base-finetuned-panx-de results: - task: name: Token Classification type: token-classification dataset: name: xtreme type: xtreme args: PAN-X.de metrics: - name: F1 type: f1 value: 0.8632527372262775 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-panx-de This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the xtreme dataset. It achieves the following results on the evaluation set: - Loss: 0.1367 - F1: 0.8633 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 24 - eval_batch_size: 24 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | |:-------------:|:-----:|:----:|:---------------:|:------:| | 0.2582 | 1.0 | 525 | 0.1653 | 0.8238 | | 0.1301 | 2.0 | 1050 | 0.1417 | 0.8439 | | 0.0841 | 3.0 | 1575 | 0.1367 | 0.8633 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.11.0+cu102 - Datasets 1.16.1 - Tokenizers 0.10.3
TeamUNIVA/Komodo_6B_v3.0.0
TeamUNIVA
"2024-03-04T11:27:20"
128
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "ko", "en", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-03-04T10:58:49"
--- license: apache-2.0 language: - ko - en --- # Base Model beomi/Yi-Ko-6B ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "TeamUNIVA/Komodo_6B_v3.0.0" model = AutoModelForCausalLM.from_pretrained(model_id) tokenizer = AutoTokenizer.from_pretrained(model_id) text = '''<|system|> 당신은 사용자의 질문에 친절하게 답변을 하는 챗봇입니다. <|user|> 안녕하세요? <|bot|> ''' inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=20) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
gavrilstep/c7c3c777-740f-4ade-ba0b-44204337981b
gavrilstep
"2025-01-25T01:30:25"
8
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "base_model:Maykeye/TinyLLama-v0", "base_model:adapter:Maykeye/TinyLLama-v0", "license:apache-2.0", "4-bit", "bitsandbytes", "region:us" ]
null
"2025-01-25T01:06:18"
--- library_name: peft license: apache-2.0 base_model: Maykeye/TinyLLama-v0 tags: - axolotl - generated_from_trainer model-index: - name: c7c3c777-740f-4ade-ba0b-44204337981b results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: Maykeye/TinyLLama-v0 bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 1b3dcfb54c6bd3d6_train_data.json ds_type: json format: custom path: /workspace/input_data/1b3dcfb54c6bd3d6_train_data.json type: field_instruction: text field_output: text_cleaned format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null device: cuda early_stopping_patience: null eval_max_new_tokens: 128 eval_steps: 5 eval_table_size: null evals_per_epoch: null flash_attention: false fp16: null gradient_accumulation_steps: 4 gradient_checkpointing: true group_by_length: false hub_model_id: gavrilstep/c7c3c777-740f-4ade-ba0b-44204337981b hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.0002 load_in_4bit: true load_in_8bit: false local_rank: null logging_steps: 3 lora_alpha: 32 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 16 lora_target_linear: true lr_scheduler: cosine max_memory: 0: 75GiB max_steps: 30 micro_batch_size: 2 mlflow_experiment_name: /tmp/1b3dcfb54c6bd3d6_train_data.json model_type: AutoModelForCausalLM num_epochs: 1 optimizer: adamw_torch output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 10 sequence_len: 1024 special_tokens: pad_token: </s> strict: false tf32: false tokenizer_type: AutoTokenizer train_on_inputs: true trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: 1ae0d02a-82ac-419b-bce9-a85da88118d7 wandb_project: Gradients-On-Demand wandb_run: your_name wandb_runid: 1ae0d02a-82ac-419b-bce9-a85da88118d7 warmup_steps: 10 weight_decay: 0.01 xformers_attention: true ``` </details><br> # c7c3c777-740f-4ade-ba0b-44204337981b This model is a fine-tuned version of [Maykeye/TinyLLama-v0](https://huggingface.co/Maykeye/TinyLLama-v0) on the None dataset. It achieves the following results on the evaluation set: - Loss: 7.3896 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - training_steps: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0000 | 1 | 7.8085 | | 7.8007 | 0.0001 | 5 | 7.7824 | | 7.8506 | 0.0002 | 10 | 7.6479 | | 7.4974 | 0.0003 | 15 | 7.5054 | | 7.3681 | 0.0004 | 20 | 7.4283 | | 7.4085 | 0.0005 | 25 | 7.3957 | | 7.627 | 0.0006 | 30 | 7.3896 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
cansuarslann/kgl
cansuarslann
"2025-02-18T17:20:20"
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:Alpha-VLLM/Lumina-Image-2.0", "base_model:adapter:Alpha-VLLM/Lumina-Image-2.0", "region:us" ]
text-to-image
"2025-02-18T17:20:17"
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: Screenshot output: url: images/Ekran Resmi 2025-02-18 20.18.42.png base_model: Alpha-VLLM/Lumina-Image-2.0 instance_prompt: null --- # abs <Gallery /> ## Download model [Download](/cansuarslann/kgl/tree/main) them in the Files & versions tab.
Omerhan/checkpoint-10934-v7
Omerhan
"2025-01-02T20:45:20"
6
0
sentence-transformers
[ "sentence-transformers", "safetensors", "xlm-roberta", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:920106", "loss:MatryoshkaLoss", "loss:MultipleNegativesRankingLoss", "tr", "arxiv:1908.10084", "arxiv:2205.13147", "arxiv:1705.00652", "base_model:intfloat/multilingual-e5-large-instruct", "base_model:finetune:intfloat/multilingual-e5-large-instruct", "license:apache-2.0", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
"2025-01-02T20:43:33"
--- language: - tr license: apache-2.0 tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:920106 - loss:MatryoshkaLoss - loss:MultipleNegativesRankingLoss base_model: intfloat/multilingual-e5-large-instruct widget: - source_sentence: Fısıh için patates cipsi kosher sentences: - 'Geçen yıl 125.000 $ kazandıysanız ve 10.000 $ düşebiliyorsanız, vergilendirilebilir geliriniz 115.000 $ ''a düşürülür. Ontario''da yaşadıysanız, indiriminiz sizi sadece 4.000 $ ''ın üzerinde vergiden kurtarır. Öte yandan, 10.000 $''lık bir vergi kredisi sadece 2,105 $ değerinde olurdu. Yukarıdaki, kesintilerin her zaman vergi kredilerinden daha iyi olduğunu gösteriyor gibi görünüyor, ancak her zaman değil: Bir vergi kredisi ve indirim, 35.000 $ vergi elde ederseniz aynı tasarrufla sonuçlanacaktır.' - Diğer şeker ikame edicilere göre daha az bir aftertate sahip olduğu iddia edilir ve fırınlanmış ürünlerde ve yumuşak içeceklerde kullanıma uygundur. Ürün Kosher - ve potansiyel olarak Hamursuz Bayramı için Kosher - ve yakın gelecekte birçok üründe görünmesini bekleyebiliriz.Ayrıca hepimiz güçlü müshiller olduklarının farkında olmalıyız!) Rosh Hashana yaklaşırken, belki de Shimshon'un benzetmesini genişletebiliriz ve bir kullanım için bir tatlı olabilir. - Hamursuz Bayramı için Kaşer işaretli patates cipsi bu sorunların hiçbirisi olmadığı için sertifikalandırılmıştır...Köşe yağında kızartılmış basit patates, Hamursuz Bayramı için Kaşer olmayan hiçbir şey yapmayan makinelerde işlenir.Fısıh Bayramı için Kaşer işaretli patates cipsleri bu sorunlardan hiçbirine sahip olmadığı için sertifikalandırılmıştır...Köşe yağında kızartılmış basit patates, Hamursuz Bayramı için Kaşer olmayan makinelerde işlenmiştir. - source_sentence: Kim söyledi mona lisa gülümsemesini kaybetti sentences: - Mona Lisa Lost Her Smile sözleri ve akorları sadece kişisel kullanımınız için tasarlanmıştır, gerçekten David Allan Coe tarafından kaydedilen güzel bir country şarkısıdır. - 'Arama Ara: Rose Müzik merkezi, Huber Heights, OH''da bulunan ve Interstate 70''in hemen dışında yer alan tamamen kapalı bir açık hava amfitiyatrosudur. Amfitiyatro, balkon koltuklarının ön sıra koltukları kadar iyi olduğu 4200 kişilik bir oturma kapasiteli mekandır. Bu tesiste nerede oturursanız oturun, bir fan olarak deneyiminizin avantajları vardır.' - Ortaya çıkan görüntüler, yüzlerce yıllık vernik ve diğer değişiklikleri ortadan kaldırıyor, sanatçının boyalı figürü nasıl hayata geçirdiğine ve da Vinci ve çağdaşlarına nasıl göründüğüne ışık tutuyor. Mona Lisa'nın yüzü biraz daha geniş görünüyor ve gülümseme farklı ve gözler farklı, dedi Cotte. - source_sentence: kovanlar bir tür gıda zehirlenmesidir sentences: - Bazen gıda zehirlenmesinden hasta hissetmek, kötü yiyecekleri yedikten sonraki saatler içinde ortaya çıkar. Diğer zamanlarda, biri birkaç gün sonraya kadar hasta hissetmeyebilir. Hafif gıda zehirlenmesi vakalarında, çok uzun süre hasta hissetmeyeceksiniz ve yakında tekrar iyi hissedeceksiniz. - Bebeklerde botulizm. genellikle kabızlığa neden olur; yetişkinlerde, ya da neden olabilir. Kabızlık veya ishal. Gıda alerjileri gıda zehirlenmesi ile karıştırılabilir. En ciddi alerjik reaksiyon türleri anidir. kaşıntı, kovanlar, nefes alma zorluğu ve düşük kan pre-. tabi. Buna anafilaksi veya alerjik şok denir. - CloseHandle. CloseHandle işlevi açık bir nesne kulpunu kapatır. BOOL CloseHandle( Handle hObject // close to close to close ; Parametreler hObject Handle to a open object. Return Values. Fonksiyon başarılı olursa, dönüş değeri sıfırdır. İşlev başarısız olursa, dönüş değeri sıfırdır. Genişletilmiş hata bilgisi almak için GetLastError. Remarks'u arayın. - source_sentence: Hint Müslüman erkek çocuk isimleri ile anlam sentences: - Hayır, hamileyseniz pişmemiş pepperoni yemek güvenli değildir. Ham gıda, listeria olarak adlandırılan zararlı bakteriler içerir. Listeria bakterileri, hamile kadınlarda beyin enfeksiyonuna ve hatta ölüme yol açabilecek listeriosis'e neden olabilir. - Bir erkek ya da kız için güzel bir isme ihtiyacınız olsun, size dünya çapında popüler isimlerin büyük bir koleksiyonunu veriyoruz. İsteğinize bağlı olarak bebeğiniz için bir Hıristiyan adı, bir Hindu adı veya bir Müslüman adı seçebilirsiniz. Bir erkek ya da kız için güzel bir isme ihtiyacınız varsa, size dünya çapında popüler isimlerin büyük bir koleksiyonunu veriyoruz. İsteğinize bağlı olarak bebeğiniz için bir Hıristiyan adı, bir Hindu adı veya bir Müslüman adı seçebilirsiniz. - '- Modern bebek erkek isimleri. - Modern bebek kız isimleri. Hint Boy ve Hint Kız İsimleri Komple Listesi. Anlamları ile bebek isimleri tam listemize göz atın, sevimli bebek fotoğrafları, anketler, zodyak etkisi ve çok daha fazlası prensesiniz veya rockstar.ee için en iyi ismi seçmek için bizim kapsamlı veritabanı popüler Hindu isimleri, benzersiz Müslüman isimleri, en iyi on Sih isimleri, A''dan Z''ye Hıristiyan isimleri, sevimli bebek Pencap isimleri, kısa ve tatlı Jain Gurati, güzel' - source_sentence: ret kuyruğu nedir sentences: - 'Bir kuyruktan gelen mesajlar ''ölü harfli'' olabilir; yani, aşağıdaki olaylardan herhangi biri meydana geldiğinde başka bir değiş tokuşa yeniden yayınlanabilir: 1 İleti, requeue=false ile (basic.reject veya basic.nack) reddedilir, 2 İletinin TTL''si sona erer; veya. 3 Kuyruk uzunluğu sınırı aşılır.' - 2.'reddetmek'. Bir fikir veya inançla aynı fikirde değilseniz,'reddetmek' demiyorsunuz. Bunu reddettiğinizi söylüyorsunuz. Bazı insanlar karma ekonomi fikrini reddediyor. Ailemin dini inançlarını reddetmek benim için zordu. 3. İsim olarak kullanılır. Reddetmek, attığınız şeylere atıfta bulunmak için kullanılan bir isimdir. - Clark County, Amerika Birleşik Devletleri'nin Wisconsin eyaletinde yer alan bir ilçedir. 2010 nüfus sayımına göre nüfusu 34.690'dır. İlçe merkezi Neillsville'dir. pipeline_tag: sentence-similarity library_name: sentence-transformers --- # intfloat-fine-tuned This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [intfloat/multilingual-e5-large-instruct](https://huggingface.co/intfloat/multilingual-e5-large-instruct) on the json dataset. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [intfloat/multilingual-e5-large-instruct](https://huggingface.co/intfloat/multilingual-e5-large-instruct) <!-- at revision c9e87c786ffac96aeaeb42863276930883923ecb --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 1024 tokens - **Similarity Function:** Cosine Similarity - **Training Dataset:** - json - **Language:** tr - **License:** apache-2.0 ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: XLMRobertaModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("Omerhan/checkpoint-10934-v7") # Run inference sentences = [ 'ret kuyruğu nedir', "Bir kuyruktan gelen mesajlar 'ölü harfli' olabilir; yani, aşağıdaki olaylardan herhangi biri meydana geldiğinde başka bir değiş tokuşa yeniden yayınlanabilir: 1 İleti, requeue=false ile (basic.reject veya basic.nack) reddedilir, 2 İletinin TTL'si sona erer; veya. 3 Kuyruk uzunluğu sınırı aşılır.", "2.'reddetmek'. Bir fikir veya inançla aynı fikirde değilseniz,'reddetmek' demiyorsunuz. Bunu reddettiğinizi söylüyorsunuz. Bazı insanlar karma ekonomi fikrini reddediyor. Ailemin dini inançlarını reddetmek benim için zordu. 3. İsim olarak kullanılır. Reddetmek, attığınız şeylere atıfta bulunmak için kullanılan bir isimdir.", ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 1024] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### json * Dataset: json * Size: 920,106 training samples * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code> * Approximate statistics based on the first 1000 samples: | | anchor | positive | negative | |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | string | | details | <ul><li>min: 4 tokens</li><li>mean: 10.38 tokens</li><li>max: 39 tokens</li></ul> | <ul><li>min: 26 tokens</li><li>mean: 81.21 tokens</li><li>max: 149 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 78.05 tokens</li><li>max: 133 tokens</li></ul> | * Samples: | anchor | positive | negative | |:----------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>Avustralya'ya özgü hangi meyve</code> | <code>Passiflora herbertiana. Avustralya'ya özgü nadir bir tutku meyvesi. Meyveler yeşil tenli, beyaz etli, bilinmeyen bir yenilebilir derecelendirmeye sahiptir. Bazı kaynaklar meyveyi yenilebilir, tatlı ve lezzetli olarak listelerken, diğerleri meyveleri acı ve yenemez olarak listeler. Avustralya'ya özgü nadir bir tutku meyvesi. Meyveler yeşil tenli, beyaz etli, bilinmeyen yenilebilir bir derecelendirmeye sahip. Bazı kaynaklar meyveyi tatlı olarak listeler.</code> | <code>Kola cevizi, Afrika'nın tropikal yağmur ormanlarına özgü bir ağaç cinsidir (Cola).</code> | | <code>meyve ağaçları türleri</code> | <code>Kiraz. Kiraz ağaçları dünya çapında bulunur. Kirazdan siyah kiraza kadar değişen 40 veya daha fazla çeşit vardır. Meyve ile birlikte, kiraz ağaçları, son derece hoş kokulu hafif ve narin pembemsi-beyaz çiçekler üretir.Omments. Submit. Mülkünüze meyve ağaçları dikmek sadece size istikrarlı bir organik meyve kaynağı sağlamakla kalmaz, aynı zamanda bahçenizi güzelleştirmenizi ve oksijeni çevreye geri vermenizi sağlar.</code> | <code>Kola cevizi, Afrika'nın tropikal yağmur ormanlarına özgü bir ağaç cinsidir (Cola).</code> | | <code>Harrison City Pa nerede yaşıyor?</code> | <code>Harrison City, Amerika Birleşik Devletleri'nin Pensilvanya eyaletinde yer alan Westmoreland County'de nüfus sayımına göre belirlenmiş bir yerdir. 2000 nüfus sayımında nüfus 155'tir.</code> | <code>En yakın şehirler: Vandling borough, PA (1.1 mil ), Simpson, PA (2.0 mil ), Union Dale borough, PA (2,1 mil ), Carbondale, PA (2,4 mil ), Waymart borough, PA (2,4 mil ), Mayfield borough, PA (2.9 mil ), Prompion borough, PA (2.9 mil ), Jermyn borough, PA (3.1 mil ).</code> | * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters: ```json { "loss": "MultipleNegativesRankingLoss", "matryoshka_dims": [ 1024 ], "matryoshka_weights": [ 1 ], "n_dims_per_step": -1 } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `gradient_accumulation_steps`: 8 - `learning_rate`: 5e-06 - `num_train_epochs`: 1 - `lr_scheduler_type`: cosine - `tf32`: True - `optim`: adamw_torch_fused - `batch_sampler`: no_duplicates #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: no - `prediction_loss_only`: True - `per_device_train_batch_size`: 8 - `per_device_eval_batch_size`: 8 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 8 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-06 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1.0 - `num_train_epochs`: 1 - `max_steps`: -1 - `lr_scheduler_type`: cosine - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: True - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch_fused - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `batch_sampler`: no_duplicates - `multi_dataset_batch_sampler`: proportional </details> ### Training Logs | Epoch | Step | Training Loss | |:------:|:-----:|:-------------:| | 0.0348 | 500 | 0.1492 | | 0.0696 | 1000 | 0.1114 | | 0.1043 | 1500 | 0.1013 | | 0.1391 | 2000 | 0.0988 | | 0.1739 | 2500 | 0.0973 | | 0.2087 | 3000 | 0.0909 | | 0.2434 | 3500 | 0.0858 | | 0.2782 | 4000 | 0.0899 | | 0.3130 | 4500 | 0.0861 | | 0.3478 | 5000 | 0.0821 | | 0.3826 | 5500 | 0.09 | | 0.4173 | 6000 | 0.078 | | 0.4521 | 6500 | 0.0796 | | 0.4869 | 7000 | 0.0816 | | 0.5217 | 7500 | 0.0867 | | 0.5565 | 8000 | 0.0787 | | 0.5912 | 8500 | 0.0691 | | 0.6260 | 9000 | 0.0755 | | 0.6608 | 9500 | 0.079 | | 0.6956 | 10000 | 0.0694 | | 0.7303 | 10500 | 0.075 | ### Framework Versions - Python: 3.10.12 - Sentence Transformers: 3.1.1 - Transformers: 4.45.2 - PyTorch: 2.5.1+cu121 - Accelerate: 1.2.1 - Datasets: 3.2.0 - Tokenizers: 0.20.3 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MatryoshkaLoss ```bibtex @misc{kusupati2024matryoshka, title={Matryoshka Representation Learning}, author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi}, year={2024}, eprint={2205.13147}, archivePrefix={arXiv}, primaryClass={cs.LG} } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
aipib/Llama-3-8B-Instruct-Ja.gguf
aipib
"2024-04-23T01:05:45"
0
2
null
[ "gguf", "endpoints_compatible", "region:us", "conversational" ]
null
"2024-04-11T12:04:41"
# Llama-3-8B-Instruct-Ja.gguf alfredplpl/Llama-3-8B-Instruct-Jaのgguf版です。取り敢えずQ3KM/Q4KMをアップロードしました。 ``` ## 💻 Usage プロンプトの最後に"<|eot_id|>" を付加して下さい。 詳細はalfredplpl/Llama-3-8B-Instruct-Jaを参照して下さい。
RamAnanth1/decision-transformers-walker2d-expert
RamAnanth1
"2022-09-29T15:40:33"
64
0
transformers
[ "transformers", "pytorch", "decision_transformer", "generated_from_trainer", "deep-reinforcement-learning", "reinforcement-learning", "decision-transformer", "gym-continous-control", "dataset:decision_transformer_gym_replay", "arxiv:2106.01345", "endpoints_compatible", "region:us" ]
reinforcement-learning
"2022-09-29T15:35:33"
--- tags: - generated_from_trainer - deep-reinforcement-learning - reinforcement-learning - decision-transformer - gym-continous-control pipeline_tag: reinforcement-learning datasets: - decision_transformer_gym_replay --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Decision Transformer model trained on expert trajectories sampled from the Gym Walker2d environment This is a trained [Decision Transformer](https://arxiv.org/abs/2106.01345) model trained from scratch on expert trajectories sampled from the Gym Walker2d environment based on the modified version of the example [training script](https://github.com/huggingface/blog/blob/main/notebooks/101_train-decision-transformers.ipynb) provided by HuggingFace ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 64 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 120 ### Training results ### Framework versions - Transformers 4.22.2 - Pytorch 1.12.1+cu113 - Datasets 2.5.1 - Tokenizers 0.12.1
shisa-ai/shisa-v1-gemma-8b
shisa-ai
"2024-05-19T18:09:06"
4
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "ja", "en", "dataset:augmxnt/ultra-orca-boros-en-ja-v1", "base_model:google/gemma-7b-it", "base_model:finetune:google/gemma-7b-it", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-05-17T16:45:38"
--- license: gemma datasets: - augmxnt/ultra-orca-boros-en-ja-v1 language: - ja - en base_model: google/gemma-7b-it --- shisa-v2 Base Model ablation Using a [fork](https://github.com/shisa-ai/shaberi) of [Lightblue's Shaberi benchmark framework](https://github.com/lightblue-tech/japanese_llm_eval): | Model | Average | ELYZA-tasks-100 | MT-Bench | Rakuda | Tengu-Bench | |----------------------------------------|---------|-----------------|----------|--------|-------------| | gpt-4-turbo-2024-04-09 | 8.75 | 8.78 | 8.74 | 9.18 | 8.31 | | CohereForAI/c4ai-command-r-plus | 7.69 | 7.50 | 7.43 | 9.05 | 6.79 | | gpt-3.5-turbo-0125 | 7.17 | 7.24 | 6.98 | 7.64 | 6.82 | | **shisa-ai/shisa-v1-llama3-70b** | **7.17**| **7.16** | **7.45** | **7.98** | **6.09** | | karakuri-ai/karakuri-lm-70b-chat-v0.1 | 6.84 | 6.86 | 6.43 | 7.85 | 6.23 | | lightblue/ao-karasu-72B | 6.81 | 7.19 | 6.54 | 7.25 | 6.27 | | **shisa-ai/shisa-v1-llama3-8b^** | **6.29**| **6.62** | **6.41** | **7.05**|**5.07** | | shisa-ai/shisa-swallowmx-13a47b-v1 | 6.17 | 6.48 | 6.07 | 7.11 | 5.03 | | **shisa-ai/shisa-v1-llama3-8b** | **6.10**| **6.52** | **6.20** | **6.37**|**5.33** | | Rakuten/RakutenAI-7B-chat | 5.58 | 5.92 | 4.60 | 6.58 | 5.24 | | shisa-ai/shisa-v1-gemma-8b | 5.64 | 6.50 | 5.42 | 5.10 | 5.55 | | augmxnt/shisa-gamma-7b-v1 | 5.56 | 5.84 | 4.00 | 6.73 | 5.68 | | lightblue/qarasu-14B-chat-plus-unleashed | 5.20 | 5.58 | 4.74 | 5.46 | 5.01 | | cyberagent/calm2-7b-chat | 4.76 | 4.90 | 3.58 | 5.75 | 4.81 | | mistralai/Mistral-7B-Instruct-v0.2 | 4.69 | 5.78 | 4.65 | 3.80 | 4.53 | | **shisa-ai/shisa-v1-yi1.5-9b** | **4.63**| **5.98** | **4.28** | **3.26**|**5.00** |
coffiee/dl166
coffiee
"2025-02-17T03:26:15"
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2025-02-17T03:25:11"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
fedecba007/SmolLM2-DPO-Final
fedecba007
"2025-01-29T10:11:40"
6
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "smol-course", "module_1", "trl", "dpo", "conversational", "arxiv:2305.18290", "base_model:fedecba007/SmolLM2-FT-MyDataset", "base_model:finetune:fedecba007/SmolLM2-FT-MyDataset", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2025-01-29T10:11:13"
--- base_model: fedecba007/SmolLM2-FT-MyDataset library_name: transformers model_name: SmolLM2-DPO-Final tags: - generated_from_trainer - smol-course - module_1 - trl - dpo licence: license --- # Model Card for SmolLM2-DPO-Final This model is a fine-tuned version of [fedecba007/SmolLM2-FT-MyDataset](https://huggingface.co/fedecba007/SmolLM2-FT-MyDataset). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="fedecba007/SmolLM2-DPO-Final", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290). ### Framework versions - TRL: 0.13.0 - Transformers: 4.47.1 - Pytorch: 2.5.1+cu121 - Datasets: 3.2.0 - Tokenizers: 0.21.0 ## Citations Cite DPO as: ```bibtex @inproceedings{rafailov2023direct, title = {{Direct Preference Optimization: Your Language Model is Secretly a Reward Model}}, author = {Rafael Rafailov and Archit Sharma and Eric Mitchell and Christopher D. Manning and Stefano Ermon and Chelsea Finn}, year = 2023, booktitle = {Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023}, url = {http://papers.nips.cc/paper_files/paper/2023/hash/a85b405ed65c6477a4fe8302b5e06ce7-Abstract-Conference.html}, editor = {Alice Oh and Tristan Naumann and Amir Globerson and Kate Saenko and Moritz Hardt and Sergey Levine}, } ``` Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
lesso03/6106c347-1b71-4070-b703-a3294aa6974f
lesso03
"2025-02-23T14:17:20"
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/CodeLlama-7b-hf-flash", "base_model:adapter:NousResearch/CodeLlama-7b-hf-flash", "region:us" ]
null
"2025-02-23T13:55:04"
--- library_name: peft base_model: NousResearch/CodeLlama-7b-hf-flash tags: - axolotl - generated_from_trainer model-index: - name: 6106c347-1b71-4070-b703-a3294aa6974f results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora auto_find_batch_size: true base_model: NousResearch/CodeLlama-7b-hf-flash bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 486ba615f54cb909_train_data.json ds_type: json format: custom path: /workspace/input_data/486ba615f54cb909_train_data.json type: field_instruction: ja_sentence field_output: en_sentence format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_max_new_tokens: 128 eval_steps: 50 evals_per_epoch: null flash_attention: false fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 2 gradient_checkpointing: false group_by_length: true hub_model_id: lesso03/6106c347-1b71-4070-b703-a3294aa6974f hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000203 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 10 lora_alpha: 32 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 16 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 500 micro_batch_size: 4 mlflow_experiment_name: /tmp/486ba615f54cb909_train_data.json model_type: AutoModelForCausalLM num_epochs: 1 optimizer: adamw_bnb_8bit output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 50 saves_per_epoch: null seed: 30 sequence_len: 512 special_tokens: pad_token: </s> strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: 574f9963-6846-4d77-b1ea-62bacab684b9 wandb_project: 03a wandb_run: your_name wandb_runid: 574f9963-6846-4d77-b1ea-62bacab684b9 warmup_steps: 50 weight_decay: 0.0 xformers_attention: null ``` </details><br> # 6106c347-1b71-4070-b703-a3294aa6974f This model is a fine-tuned version of [NousResearch/CodeLlama-7b-hf-flash](https://huggingface.co/NousResearch/CodeLlama-7b-hf-flash) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2389 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000203 - train_batch_size: 4 - eval_batch_size: 4 - seed: 30 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 50 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0004 | 1 | 3.2889 | | 3.2344 | 0.0176 | 50 | 1.6177 | | 2.7459 | 0.0353 | 100 | 1.4442 | | 2.565 | 0.0529 | 150 | 1.3767 | | 2.3303 | 0.0705 | 200 | 1.3169 | | 2.2378 | 0.0882 | 250 | 1.2866 | | 2.2343 | 0.1058 | 300 | 1.2735 | | 2.2444 | 0.1235 | 350 | 1.2538 | | 2.2407 | 0.1411 | 400 | 1.2466 | | 2.101 | 0.1587 | 450 | 1.2398 | | 1.9032 | 0.1764 | 500 | 1.2389 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
Nitipoom/matcha888
Nitipoom
"2025-02-12T20:41:09"
0
0
null
[ "text-to-audio", "base_model:deepseek-ai/DeepSeek-R1", "base_model:finetune:deepseek-ai/DeepSeek-R1", "license:llama3.3", "region:us" ]
text-to-audio
"2025-02-12T20:39:04"
--- license: llama3.3 base_model: - deepseek-ai/DeepSeek-R1 new_version: deepseek-ai/DeepSeek-R1 pipeline_tag: text-to-audio ---
vladjr/roberta-xlm-teste
vladjr
"2023-10-12T14:24:41"
3
0
transformers
[ "transformers", "tf", "xlm-roberta", "text-classification", "generated_from_keras_callback", "base_model:FacebookAI/xlm-roberta-base", "base_model:finetune:FacebookAI/xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
"2023-10-12T01:40:26"
--- license: mit base_model: xlm-roberta-base tags: - generated_from_keras_callback model-index: - name: vladjr/roberta-xlm-teste results: [] --- <!-- This model card has been generated automatically according to the information Keras had access to. You should probably proofread and complete it, then remove this comment. --> # vladjr/roberta-xlm-teste This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Train Loss: 0.3022 - Validation Loss: 0.1713 - Train Accuracy: 0.9577 - Epoch: 2 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 2100, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False} - training_precision: float32 ### Training results | Train Loss | Validation Loss | Train Accuracy | Epoch | |:----------:|:---------------:|:--------------:|:-----:| | 2.4904 | 0.8741 | 0.8810 | 0 | | 0.6902 | 0.2997 | 0.9375 | 1 | | 0.3022 | 0.1713 | 0.9577 | 2 | ### Framework versions - Transformers 4.34.0 - TensorFlow 2.13.0 - Datasets 2.14.5 - Tokenizers 0.14.1
ZeroAgency/o1_t-lite-it-1.0_lora
ZeroAgency
"2025-01-04T13:00:52"
5
4
peft
[ "peft", "chat", "o1", "cot", "thinking", "reflection", "question-answering", "ru", "en", "dataset:Egor-AI/Russian_thinking_dataset", "base_model:t-tech/T-lite-it-1.0", "base_model:adapter:t-tech/T-lite-it-1.0", "license:mit", "region:us" ]
question-answering
"2025-01-04T12:51:15"
--- license: mit datasets: - Egor-AI/Russian_thinking_dataset language: - ru - en base_model: - t-tech/T-lite-it-1.0 pipeline_tag: question-answering library_name: peft tags: - chat - o1 - cot - thinking - reflection --- # Russian o1 / T-lite-it-1.0 LoRA Based on https://huggingface.co/evilfreelancer/o1_t-lite-it-1.0_lora LoRA-адаптер для модели [T-lite-it-1.0](https://huggingface.co/t-tech/T-lite-it-1.0) обученный на датасете [Egor-AI/Russian_thinking_dataset](https://huggingface.co/datasets/Egor-AI/Russian_thinking_dataset) (машинный перевод на русский язык датасета [BintangFortuna/OpenO1-SFT-EN-SY](https://huggingface.co/datasets/BintangFortuna/OpenO1-SFT-EN-SY)). Обученная модель способна имитировать логические размышлению на русском языке по аналогии с тем, как это делает `o1` от `OpenAI`. Необходимо использовать следующего вида системный промт: ``` Вы — ИИ-помощник. Отформатируйте свои ответы следующим образом: <Thought> Ваши мысли (понимание, рассуждения) </Thought> <output> Ваш ответ </output> ``` Обучение производилось при помощи утилиты [impruver](https://github.com/EvilFreelancer/impruver) используя конфигурацию [T-lite-it/7B_lora_thinking](https://github.com/EvilFreelancer/impruver/blob/main/recipes/configs/T-lite-it/7B_lora_thinking.yaml) с донастройкой: - load_in_4bit: false - без max_tokens_count - optim: adamw_8bit - gradient_accumulation_steps: 1 На всё про всё ушло примерно 17.6 часов на 1xH100 80GB, при этом понадобилось 67Гб видеопамяти. Результатирующий eval_loss: 0.5200754404067993 W&B run: https://wandb.ai/b37h3z3n/trains/runs/6vwvuu46?nw=nwuserb37h3z3n ```yaml output_dir: ./models/T-lite-it_7B_lora_thinking train_path: ./train.T-lite-it_7B_lora_thinking.jsonl val_path: ./val.T-lite-it_7B_lora_thinking.jsonl datasets: - name: Egor-AI/Russian_thinking_dataset converter: impruver.instruction_to_messages add_global_bos: false add_global_eos: false mapping: system: system instruction: prompt output: response model: class: transformers.AutoModelForCausalLM name: t-tech/T-lite-it-1.0 load_in_4bit: false load_in_8bit: false dtype: bf16 lora: r: 16 lora_alpha: 16 lora_dropout: 0 bias: none target_modules: [ q_proj, k_proj, v_proj, o_proj, gate_proj, down_proj, up_proj ] task_type: CAUSAL_LM tokenizer: class: transformers.AutoTokenizer name: t-tech/T-lite-it-1.0 # max_tokens_count: 1500 trainer: eval_strategy: steps save_strategy: steps eval_steps: 100 save_steps: 100 per_device_train_batch_size: 1 per_device_eval_batch_size: 1 gradient_accumulation_steps: 1 logging_steps: 10 learning_rate: 0.0004 num_train_epochs: 3 lr_scheduler_type: cosine warmup_steps: 16 optim: adamw_8bit metric_for_best_model: eval_loss load_best_model_at_end: true save_total_limit: 2 seed: 42 remove_unused_columns: false max_grad_norm: 1.0 weight_decay: 0.08 torch_compile: false ```
aaliyaan/t5-small-finetuned-career
aaliyaan
"2024-12-08T15:16:09"
114
0
transformers
[ "transformers", "safetensors", "t5", "text2text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
"2024-12-08T15:15:51"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
gabski/deberta-claim-improvement-suggestion
gabski
"2023-06-08T17:03:54"
126
0
transformers
[ "transformers", "pytorch", "deberta", "text-classification", "en", "dataset:ClaimRev", "arxiv:2305.16799", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
"2023-06-08T14:27:26"
--- license: cc-by-nc-sa-4.0 language: - en library_name: transformers pipeline_tag: text-classification datasets: - ClaimRev widget: - text: "Teachers are likely to educate children better than parents." --- # Model This model was obtained by fine-tuning `microsoft/deberta-base` on the extended ClaimRev dataset. Paper: [To Revise or Not to Revise: Learning to Detect Improvable Claims for Argumentative Writing Support](https://arxiv.org/abs/2305.16799) Authors: Gabriella Skitalinskaya and Henning Wachsmuth # Claim Improvement Suggestion We cast this task as a multi-class classification task, where the objective is given an argumentative claim, select all types of quality issues from a defined set that should be improved when revising the claim. # Usage ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch tokenizer = AutoTokenizer.from_pretrained("gabski/deberta-claim-improvement-suggestion") model = AutoModelForSequenceClassification.from_pretrained("gabski/deberta-claim-improvement-suggestion") claim = 'Teachers are likely to educate children better than parents.' model_input = tokenizer(claim, return_tensors='pt') model_outputs = model(**model_input) outputs = torch.nn.functional.softmax(model_outputs.logits, dim = -1) print(outputs) ```
mahesh006/ai_imagesVsoriginal_images
mahesh006
"2023-03-20T17:51:46"
0
0
fastai
[ "fastai", "region:us" ]
null
"2023-03-20T17:51:22"
--- tags: - fastai --- # Amazing! 🥳 Congratulations on hosting your fastai model on the Hugging Face Hub! # Some next steps 1. Fill out this model card with more information (see the template below and the [documentation here](https://huggingface.co/docs/hub/model-repos))! 2. Create a demo in Gradio or Streamlit using 🤗 Spaces ([documentation here](https://huggingface.co/docs/hub/spaces)). 3. Join the fastai community on the [Fastai Discord](https://discord.com/invite/YKrxeNn)! Greetings fellow fastlearner 🤝! Don't forget to delete this content from your model card. --- # Model card ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed
LarkAI/bart_large_nl2sql
LarkAI
"2023-05-23T09:42:59"
118
3
transformers
[ "transformers", "pytorch", "bart", "feature-extraction", "nl2sql", "text2text-generation", "en", "dataset:wikisql", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text2text-generation
"2023-05-19T09:45:29"
--- license: apache-2.0 datasets: - wikisql language: - en pipeline_tag: text2text-generation tags: - nl2sql widget: - text: "question: get people name with age less 25 table: id, name, age" example_title: "less than" - text: "question: get people name with age equal 25 table: id, name, age" example_title: "equal" --- new version: [LarkAI/codet5p-770m_nl2sql_oig](https://huggingface.co/LarkAI/codet5p-770m_nl2sql_oig) use oig-sql dataset and support more complex sql parse # How to Use ```python import torch from transformers import AutoTokenizer, BartForConditionalGeneration device = torch.device('cuda:0') tokenizer = AutoTokenizer.from_pretrained("LarkAI/bart_large_nl2sql") model = BartForConditionalGeneration.from_pretrained("LarkAI/bart_large_nl2sql").to(device) text = "question: get people name with age less 25 table: id, name, age" inputs = tokenizer([text], max_length=1024, return_tensors="pt") output_ids = model.generate(inputs["input_ids"].to(device), num_beams=self.beams, max_length=128, min_length=8) response_text = tokenizer.batch_decode(output_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] # SELECT name FROM table WHERE age < 25 ``` reference: [juierror/flan-t5-text2sql-with-schema](https://huggingface.co/juierror/flan-t5-text2sql-with-schema) - fix this [discussion](https://huggingface.co/juierror/flan-t5-text2sql-with-schema/discussions/5) # How to Train Quick start: https://github.com/huggingface/transformers/blob/main/examples/pytorch/summarization/README.md
Mozilla/distilbert-uncased-NER-LoRA
Mozilla
"2024-12-10T16:55:28"
124
0
transformers
[ "transformers", "onnx", "safetensors", "distilbert", "token-classification", "en", "dataset:Mozilla/smart_ner_dataset", "base_model:distilbert/distilbert-base-uncased", "base_model:quantized:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
"2024-10-10T22:08:21"
--- library_name: transformers license: apache-2.0 datasets: - Mozilla/smart_ner_dataset language: - en base_model: - distilbert/distilbert-base-uncased pipeline_tag: token-classification --- # Fine-tuned LoRA Token classification on distilbert This is a fine-tuned LoRA token classifier on distilbert, designed for NER on multiple categories PERSON, ORG, CITY, STATE, CITY_STATE. ## Model Details ### Model Description This model is based on distilbert/distilbert-base-uncased and fine-tuned using LoRA for token classification. The fine-tuning process adapts the model to predict tokens across 10 categories: "O" # Outside any named entity "B-PER" # Beginning of a person entity "I-PER" # Inside a person entity "B-ORG" # Beginning of an organization entity "I-ORG" # Inside an organization entity "B-CITY" # Beginning of a city entity "I-CITY" # Inside a city entity "B-STATE" # Beginning of a state entity "I-STATE" # Inside a state entity "B-CITYSTATE" # Beginning of a city_state entity "I-CITYSTATE" # Inside a city_state entity --- - **Developed by:** Mozilla - **Language(s):** English (`en`) - **License:** Apache-2.0 - **Fine-tuned from:** [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) ### Model Sources - **Repository:** [Mozilla Smart Intent Project](https://github.com/mozilla/smart_intent) ### Citation If you use this model, please cite it as: ``` @misc{mozilla_distilbert_lora_ner, title = {Fine-tuned LoRA Token Classifier on DistilBERT}, author = {Mozilla}, year = {2024}, url = {https://huggingface.co/Mozilla/distilbert-finetuned-LoRA-token-classifier}, license = {Apache-2.0} } ```
lesso16/7bfdd24e-f3e3-4282-8d22-9a0ce6a453c4
lesso16
"2025-02-18T21:48:02"
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "base_model:scb10x/llama-3-typhoon-v1.5-8b-instruct", "base_model:adapter:scb10x/llama-3-typhoon-v1.5-8b-instruct", "license:llama3", "region:us" ]
null
"2025-02-18T19:33:46"
--- library_name: peft license: llama3 base_model: scb10x/llama-3-typhoon-v1.5-8b-instruct tags: - axolotl - generated_from_trainer model-index: - name: 7bfdd24e-f3e3-4282-8d22-9a0ce6a453c4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <br> # 7bfdd24e-f3e3-4282-8d22-9a0ce6a453c4 This model is a fine-tuned version of [scb10x/llama-3-typhoon-v1.5-8b-instruct](https://huggingface.co/scb10x/llama-3-typhoon-v1.5-8b-instruct) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.1637 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000216 - train_batch_size: 4 - eval_batch_size: 4 - seed: 160 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 50 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0001 | 1 | 3.0896 | | 2.281 | 0.0028 | 50 | 2.6474 | | 2.1115 | 0.0057 | 100 | 2.5062 | | 1.8471 | 0.0085 | 150 | 2.4059 | | 1.996 | 0.0113 | 200 | 2.3076 | | 1.991 | 0.0141 | 250 | 2.2510 | | 1.9247 | 0.0170 | 300 | 2.2153 | | 1.8468 | 0.0198 | 350 | 2.1959 | | 1.8863 | 0.0226 | 400 | 2.1697 | | 1.8793 | 0.0255 | 450 | 2.1657 | | 1.8291 | 0.0283 | 500 | 2.1637 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
JaqueMate/shuffle_3b_checkpoint-600
JaqueMate
"2025-02-12T11:26:31"
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:meta-llama/Llama-3.2-3B-Instruct", "base_model:adapter:meta-llama/Llama-3.2-3B-Instruct", "region:us" ]
null
"2025-02-12T11:26:04"
--- base_model: meta-llama/Llama-3.2-3B-Instruct library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.12.1.dev0
ChenMnZ/Llama-3-70b-EfficientQAT-w2g64
ChenMnZ
"2024-07-22T02:44:25"
9
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:2407.11062", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-07-03T12:20:54"
# EfficientQAT [EfficientQAT](https://arxiv.org/abs/2407.11062) is a novel quantization technical, which pushes the limitation of uniform (INT) quantization in an efficient manner. Due to the leverage of standard INT quantization, the quantized model of EfficientQAT can also be transferred into other formats, such as GPTQ, BitBLAS, etc. In this repo, we provide three type checkpoints, one is EQAT, indicats the original checkpoints of EfficientQAT. The other two are GPTQ and BitBLAS respectively. ## Model Zoo We provide a number of prequantized EfficientQAT models as follows: - WikiText2 PPL is measured in 2048 context length. - Avg. Accuracy indicate the average accuracy in 5 zero-shot reasoning tasks (WinoGrande,PIQA,HellaSwag,Arc-Easy, Arc-Challenge) with [lm-eval v0.4.2](https://github.com/EleutherAI/lm-evaluation-harness). - 1GB = $10^9$ Bit - Hub Link: EQAT indicates the original checkpoints. We also transfer the checkpoints into GPTQ and BitBLAS formats, which can be loaded directly through [GPTQModel](https://github.com/ModelCloud/GPTQModel). (PS: [GPTQModel](https://github.com/ModelCloud/GPTQModel) is a official bug-fixed repo of AutoGPTQ, which would be merged into [AutoGPTQ](https://github.com/AutoGPTQ/AutoGPTQ) in future.) | Model | Quantization | WikiText2 PPL | Avg. Accuracy | Model Size (GB) | Hub link| |-------|--------------|---------------|---------------|-----------------|----------| Llama-2-7B|fp16|5.47|64.86|13.2|-| Llama-2-7B|w4g128|5.53|64.27|3.7|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-2-7b-EfficientQAT-w4g128-BitBLAS)| Llama-2-7B|w3g128|5.81|64.02|3.1|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w3g128)| Llama-2-7B|w2g64|6.86|60.14|2.3|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w2g64-GPTQ)\|[BitBLAS](Llama-2-7b-EfficientQAT-w2g64-BitBLAS)| Llama-2-7B|w2g128|7.17|59.50|2.2|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-2-7b-EfficientQAT-w2g128-BitBLAS)| Llama-2-13B|fp16|4.88|67.81|25.4|-| Llama-2-13B|w4g128|4.93|67.52|6.8|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-7b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-2-7b-EfficientQAT-w4g128-BitBLAS)| Llama-2-13B|w3g128|5.12|67.28|5.6|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w3g128)| Llama-2-13B|w2g64|5.96|64.88|4.0|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w2g64-GPTQ)\|[BitBLAS](Llama-2-13b-EfficientQAT-w2g64-BitBLAS)| Llama-2-13B|w2g128|6.08|63.88|3.8|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-13b-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-2-13b-EfficientQAT-w2g128-BitBLAS)| Llama-2-70B|fp16|3.32|72.41|131.6|-| Llama-2-70B|w4g128|3.39|72.62|35.8|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-2-70b-EfficientQAT-w4g128-BitBLAS)| Llama-2-70B|w3g128|3.61|71.76|29.1|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w3g128)| Llama-2-70B|w2g64|4.52|69.48|20.1|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w2g64-GPTQ)\|[BitBLAS](Llama-2-70b-EfficientQAT-w2g64-BitBLAS)| Llama-2-70B|w2g128|4.61|68.93|18.9|[EQAT](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-2-70b-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-2-70b-EfficientQAT-w2g128-BitBLAS)| Llama-3-8B|fp16|6.14|68.58|13.0|-| Llama-3-8B|w4g128|6.47|68.43|5.4|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-3-8b-EfficientQAT-w4g128-BitBLAS)| Llama-3-8B|w3g128|7.09|67.35|4.7|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w3g128)| Llama-3-8B|w2g64|9.41|60.76|3.9|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-3-8b-EfficientQAT-w2g64-BitBLAS)| Llama-3-8B|w2g128|9.80|59.36|3.8|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-3-8b-EfficientQAT-w2g128-BitBLAS)| Llama-3-70B|fp16|2.85|75.33|137.8|-| Llama-3-70B|w4g128|3.17|74.57|38.9|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-3-70b-EfficientQAT-w4g128-BitBLAS)| Llama-3-70B|w3g128|4.19|72.42|32.2|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w3g128)| Llama-3-70B|w2g64|6.08|67.89|23.2|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w2g64-GPTQ)| Llama-3-70B|w2g128|6.38|67.57|22.0|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-3-70b-EfficientQAT-w2g128-BitBLAS)| Llama-3-8B-Instruct|fp16|8.29|68.43|13.0|-| Llama-3-8B-Instruct|w4g128|7.93|68.39|5.4|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-3-8b-instruct-EfficientQAT-w4g128-BitBLAS)| Llama-3-8B-Instruct|w3g128|8.55|67.24|4.7|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w3g128)| Llama-3-8B-Instruct|w2g64|11.19|60.66|3.9|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g64-GPTQ)\|[BitBLAS](Llama-3-8b-instruct-EfficientQAT-w2g64-BitBLAS)| Llama-3-8B-Instruct|w2g128|11.73|60.16|3.8|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-8b-instruct-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-3-8b-instruct-EfficientQAT-w2g128-BitBLAS)| Llama-3-70B-Instruct|fp16|5.33|73.78|137.8|-| Llama-3-70B-Instruct|w4g128|5.35|73.47|38.9|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w4g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w4g128-GPTQ)\|[BitBLAS](Llama-3-70b-instruct-EfficientQAT-w4g128-BitBLAS)| Llama-3-70B-Instruct|w3g128|5.65|72.87|32.2|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w3g128)| Llama-3-70B-Instruct|w2g64|7.86|67.64|23.2|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w2g64)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w2g64-GPTQ)\|[BitBLAS](Llama-3-70b-instruct-EfficientQAT-w2g64-BitBLAS)| Llama-3-70B-Instruct|w2g128|8.14|67.54|22.0|[EQAT](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w2g128)\|[GPTQ](https://huggingface.co/ChenMnZ/Llama-3-70b-instruct-EfficientQAT-w2g128-GPTQ)\|[BitBLAS](Llama-3-70b-instruct-EfficientQAT-w2g128-BitBLAS)| ## Usage of EQAT models Please refer [https://github.com/OpenGVLab/EfficientQAT](https://github.com/OpenGVLab/EfficientQAT?tab=readme-ov-file#inference) for details. ## Usage of GPTQ and BitBLAS models Below is an example to inference with GPTQ or BitBLAS quantized formats. ```Python from transformers import AutoTokenizer from gptqmodel import GPTQModel quant_dir = "ChenMnZ/Llama-2-7b-EfficientQAT-w2g128-GPTQ" # quant_dir = "ChenMnZ/Llama-2-7b-EfficientQAT-w2g128-BitBLAS" # or local path tokenizer = AutoTokenizer.from_pretrained(quant_dir, use_fast=True) # load quantized model to the first GPU model = GPTQModel.from_quantized(quant_dir) # inference with model.generate print(tokenizer.decode(model.generate(**tokenizer("Model quantization is", return_tensors="pt").to(model.device))[0])) ``` ## Citation If you found this work useful, please consider citing: ``` @article{efficientqat, title={EfficientQAT: Efficient Quantization-Aware Training for Large Language Models}, author={Chen, Mengzhao and Shao, Wenqi and Xu, Peng and Wang, Jiahao and Gao, Peng and Zhang, Kaipeng and Qiao, Yu and Luo, Ping}, journal={arXiv preprint arXiv:2407.11062}, year={2024} } ```
MoEMoEKKung/Frankenstein-MoE-en-10.7Bx4
MoEMoEKKung
"2024-01-06T13:24:39"
1,532
0
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "en", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-01-06T09:20:22"
--- language: - en license: cc-by-nc-sa-4.0 --- # Frankenstein-MoE ### Method To initialize the gate projection weight of the MoE layer, the H6 trainset was sampled and used. We sampled 400 and selected the final 30 with low PPL. trufulqa used gpt4 to generate data. ### Evals in progress
fine-tuned/BAAI_bge-small-en-v1_5-05062024-x987-webapp
fine-tuned
"2024-06-05T11:31:45"
7
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "feature-extraction", "sentence-similarity", "mteb", "Argument", "Counter", "Discussion", "Perspective", "Opposing", "en", "dataset:fine-tuned/BAAI_bge-small-en-v1_5-05062024-x987-webapp", "dataset:allenai/c4", "license:apache-2.0", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
"2024-06-05T11:31:40"
--- license: apache-2.0 datasets: - fine-tuned/BAAI_bge-small-en-v1_5-05062024-x987-webapp - allenai/c4 language: - en pipeline_tag: feature-extraction tags: - sentence-transformers - feature-extraction - sentence-similarity - mteb - Argument - Counter - Discussion - Perspective - Opposing --- This model is a fine-tuned version of [**BAAI/bge-small-en-v1.5**](https://huggingface.co/BAAI/bge-small-en-v1.5) designed for the following use case: debate ## How to Use This model can be easily integrated into your NLP pipeline for tasks such as text classification, sentiment analysis, entity recognition, and more. Here's a simple example to get you started: ```python from sentence_transformers import SentenceTransformer from sentence_transformers.util import cos_sim model = SentenceTransformer( 'fine-tuned/BAAI_bge-small-en-v1_5-05062024-x987-webapp', trust_remote_code=True ) embeddings = model.encode([ 'first text to embed', 'second text to embed' ]) print(cos_sim(embeddings[0], embeddings[1])) ```
SubhasishSaha/dqn-LunarLander-v2
SubhasishSaha
"2024-03-21T05:17:17"
2
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
"2024-03-21T04:36:05"
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: -446.72 +/- 136.10 name: mean_reward verified: false --- # **DQN** Agent playing **LunarLander-v2** This is a trained model of a **DQN** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
fifxus/7898af8d-702c-4555-9ded-4863503d0e60
fifxus
"2025-02-03T10:33:57"
11
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/Yarn-Solar-10b-64k", "base_model:adapter:NousResearch/Yarn-Solar-10b-64k", "license:apache-2.0", "8-bit", "bitsandbytes", "region:us" ]
null
"2025-02-03T09:26:45"
--- library_name: peft license: apache-2.0 base_model: NousResearch/Yarn-Solar-10b-64k tags: - axolotl - generated_from_trainer model-index: - name: 7898af8d-702c-4555-9ded-4863503d0e60 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: NousResearch/Yarn-Solar-10b-64k bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 0ff2655235b8a4c4_train_data.json ds_type: json format: custom path: /workspace/input_data/0ff2655235b8a4c4_train_data.json type: field_instruction: ja field_output: en format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null device_map: auto do_eval: true early_stopping_patience: null eval_batch_size: 2 eval_max_new_tokens: 128 eval_steps: null eval_table_size: null evals_per_epoch: null flash_attention: true fp16: null fsdp: null fsdp_config: null gradient_accumulation_steps: 4 gradient_checkpointing: true gradient_clipping: 1.0 group_by_length: true hub_model_id: fifxus/7898af8d-702c-4555-9ded-4863503d0e60 hub_repo: null hub_strategy: end hub_token: null learning_rate: 0.0001 load_in_4bit: true load_in_8bit: true local_rank: null logging_steps: 1 lora_alpha: 16 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 8 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_memory: 0: 75GB max_steps: 200 micro_batch_size: 2 mlflow_experiment_name: /tmp/0ff2655235b8a4c4_train_data.json model_type: AutoModelForCausalLM num_epochs: 1 optimizer: adamw_bnb_8bit output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: null saves_per_epoch: null sequence_len: 1024 special_tokens: pad_token: </s> strict: false tf32: false tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: techspear-hub wandb_mode: online wandb_name: a303f412-1c67-4884-af8c-e465e1e21698 wandb_project: Gradients-On-10 wandb_run: your_name wandb_runid: a303f412-1c67-4884-af8c-e465e1e21698 warmup_steps: 5 weight_decay: 0.01 xformers_attention: null ``` </details><br> # 7898af8d-702c-4555-9ded-4863503d0e60 This model is a fine-tuned version of [NousResearch/Yarn-Solar-10b-64k](https://huggingface.co/NousResearch/Yarn-Solar-10b-64k) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.7074 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 5 - training_steps: 200 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 2.5548 | 0.0114 | 200 | 0.7074 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
harish/t5-e2e-10epochs-lr1e4-alpha0-9
harish
"2022-08-11T14:15:14"
5
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
"2022-08-11T14:10:36"
--- license: cc-by-nc-sa-4.0 ---
juliensimon/autonlp-song-lyrics-18753423
juliensimon
"2021-10-15T09:55:11"
5
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "autonlp", "en", "dataset:juliensimon/autonlp-data-song-lyrics", "co2_eq_emissions", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
"2022-03-02T23:29:05"
--- tags: autonlp language: en widget: - text: "I love AutoNLP 🤗" datasets: - juliensimon/autonlp-data-song-lyrics co2_eq_emissions: 55.552987716859484 --- # Model Trained Using AutoNLP - Problem type: Multi-class Classification - Model ID: 18753423 - CO2 Emissions (in grams): 55.552987716859484 ## Validation Metrics - Loss: 0.913820743560791 - Accuracy: 0.654110224531453 - Macro F1: 0.5327761649415296 - Micro F1: 0.654110224531453 - Weighted F1: 0.6339481529454227 - Macro Precision: 0.6799297267808116 - Micro Precision: 0.654110224531453 - Weighted Precision: 0.6533459269990771 - Macro Recall: 0.49907494605289154 - Micro Recall: 0.654110224531453 - Weighted Recall: 0.654110224531453 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/juliensimon/autonlp-song-lyrics-18753423 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("juliensimon/autonlp-song-lyrics-18753423", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("juliensimon/autonlp-song-lyrics-18753423", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
davera-017/Reinforce-Pixelcopter-PLE-v2
davera-017
"2023-09-13T20:32:17"
0
0
null
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
"2023-09-13T20:12:39"
--- tags: - Pixelcopter-PLE-v0 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-Pixelcopter-PLE-v2 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Pixelcopter-PLE-v0 type: Pixelcopter-PLE-v0 metrics: - type: mean_reward value: 5.20 +/- 9.09 name: mean_reward verified: false --- # **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
ferrazzipietro/Llama-2-13b-chat-hf__adapters_en.layer1_4_torch.bfloat16_32_32_0.05_8_0.0002
ferrazzipietro
"2024-03-11T14:34:11"
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
"2024-03-11T14:33:36"
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
sd-concepts-library/malika-favre-art-style
sd-concepts-library
"2022-09-08T19:08:58"
0
28
null
[ "license:mit", "region:us" ]
null
"2022-09-08T19:08:53"
--- license: mit --- ### Malika Favre Art Style on Stable Diffusion This is the `<malika-favre>` concept taught to Stable Diffusion via Textual Inversion. You can load this concept into the [Stable Conceptualizer](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/stable_conceptualizer_inference.ipynb) notebook. You can also train your own concepts and load them into the concept libraries using [this notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_textual_inversion_training.ipynb). Here is the new concept you will be able to use as a `style`: ![<malika-favre> 0](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/9.jpeg) ![<malika-favre> 1](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/5.jpeg) ![<malika-favre> 2](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/12.jpeg) ![<malika-favre> 3](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/2.jpeg) ![<malika-favre> 4](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/6.jpeg) ![<malika-favre> 5](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/7.jpeg) ![<malika-favre> 6](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/8.jpeg) ![<malika-favre> 7](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/0.jpeg) ![<malika-favre> 8](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/1.jpeg) ![<malika-favre> 9](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/10.jpeg) ![<malika-favre> 10](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/11.jpeg) ![<malika-favre> 11](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/4.jpeg) ![<malika-favre> 12](https://huggingface.co/sd-concepts-library/malika-favre-art-style/resolve/main/concept_images/3.jpeg)
JFernandoGRE/bert_mlm_SALAS_VALDIVIA_MARGARITA_ELIZABETH
JFernandoGRE
"2024-08-12T07:19:59"
5
0
null
[ "safetensors", "bert", "generated_from_trainer", "base_model:dccuchile/bert-base-spanish-wwm-uncased", "base_model:finetune:dccuchile/bert-base-spanish-wwm-uncased", "region:us" ]
null
"2024-08-08T10:47:30"
--- base_model: dccuchile/bert-base-spanish-wwm-uncased tags: - generated_from_trainer model-index: - name: bert_mlm_SALAS_VALDIVIA_MARGARITA_ELIZABETH results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert_mlm_SALAS_VALDIVIA_MARGARITA_ELIZABETH This model is a fine-tuned version of [dccuchile/bert-base-spanish-wwm-uncased](https://huggingface.co/dccuchile/bert-base-spanish-wwm-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.3439 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.6782 | 1.0 | 96 | 2.8701 | | 2.7894 | 2.0 | 192 | 2.4178 | | 2.5276 | 3.0 | 288 | 2.2732 | ### Framework versions - Transformers 4.43.3 - Pytorch 2.4.0+cu121 - Datasets 2.20.0 - Tokenizers 0.19.1
dang1812/trained_model
dang1812
"2023-04-24T03:42:41"
108
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
"2023-04-24T03:26:47"
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: trained_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # trained_model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.0+cu118 - Datasets 2.11.0 - Tokenizers 0.13.3
suiyanjun/lora_model2
suiyanjun
"2025-02-19T01:42:26"
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
"2025-02-19T01:42:16"
--- base_model: unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** suiyanjun - **License:** apache-2.0 - **Finetuned from model :** unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
davidschulte/ESM_nguha__legalbench_contract_nli_explicit_identification
davidschulte
"2024-12-09T22:26:40"
7
0
null
[ "safetensors", "embedding_space_map", "BaseLM:bert-base-multilingual-uncased", "dataset:nguha/legalbench", "arxiv:2410.15148", "base_model:google-bert/bert-base-multilingual-uncased", "base_model:finetune:google-bert/bert-base-multilingual-uncased", "license:apache-2.0", "region:us" ]
null
"2024-12-09T22:26:35"
--- base_model: bert-base-multilingual-uncased datasets: - nguha/legalbench license: apache-2.0 tags: - embedding_space_map - BaseLM:bert-base-multilingual-uncased --- # ESM nguha/legalbench <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> ESM - **Developed by:** David Schulte - **Model type:** ESM - **Base Model:** bert-base-multilingual-uncased - **Intermediate Task:** nguha/legalbench - **ESM architecture:** linear - **Language(s) (NLP):** [More Information Needed] - **License:** Apache-2.0 license ## Training Details ### Intermediate Task - **Task ID:** nguha/legalbench - **Subset [optional]:** contract_nli_explicit_identification - **Text Column:** text - **Label Column:** answer - **Dataset Split:** train - **Sample size [optional]:** 8 - **Sample seed [optional]:** ### Training Procedure [optional] <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Language Model Training Hyperparameters [optional] - **Epochs:** 3 - **Batch size:** 32 - **Learning rate:** 2e-05 - **Weight Decay:** 0.01 - **Optimizer**: AdamW ### ESM Training Hyperparameters [optional] - **Epochs:** 10 - **Batch size:** 32 - **Learning rate:** 0.001 - **Weight Decay:** 0.01 - **Optimizer**: AdamW ### Additional trainiung details [optional] ## Model evaluation ### Evaluation of fine-tuned language model [optional] ### Evaluation of ESM [optional] MSE: ### Additional evaluation details [optional] ## What are Embedding Space Maps? <!-- This section describes the evaluation protocols and provides the results. --> Embedding Space Maps (ESMs) are neural networks that approximate the effect of fine-tuning a language model on a task. They can be used to quickly transform embeddings from a base model to approximate how a fine-tuned model would embed the the input text. ESMs can be used for intermediate task selection with the ESM-LogME workflow. ## How can I use Embedding Space Maps for Intermediate Task Selection? [![PyPI version](https://img.shields.io/pypi/v/hf-dataset-selector.svg)](https://pypi.org/project/hf-dataset-selector) We release **hf-dataset-selector**, a Python package for intermediate task selection using Embedding Space Maps. **hf-dataset-selector** fetches ESMs for a given language model and uses it to find the best dataset for applying intermediate training to the target task. ESMs are found by their tags on the Huggingface Hub. ```python from hfselect import Dataset, compute_task_ranking # Load target dataset from the Hugging Face Hub dataset = Dataset.from_hugging_face( name="stanfordnlp/imdb", split="train", text_col="text", label_col="label", is_regression=False, num_examples=1000, seed=42 ) # Fetch ESMs and rank tasks task_ranking = compute_task_ranking( dataset=dataset, model_name="bert-base-multilingual-uncased" ) # Display top 5 recommendations print(task_ranking[:5]) ``` For more information on how to use ESMs please have a look at the [official Github repository](https://github.com/davidschulte/hf-dataset-selector). ## Citation <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> If you are using this Embedding Space Maps, please cite our [paper](https://arxiv.org/abs/2410.15148). **BibTeX:** ``` @misc{schulte2024moreparameterefficientselectionintermediate, title={Less is More: Parameter-Efficient Selection of Intermediate Tasks for Transfer Learning}, author={David Schulte and Felix Hamborg and Alan Akbik}, year={2024}, eprint={2410.15148}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2410.15148}, } ``` **APA:** ``` Schulte, D., Hamborg, F., & Akbik, A. (2024). Less is More: Parameter-Efficient Selection of Intermediate Tasks for Transfer Learning. arXiv preprint arXiv:2410.15148. ``` ## Additional Information
Splend1dchan/wav2vec2-large-10min-lv60-self
Splend1dchan
"2022-05-30T04:37:27"
5
0
transformers
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "speech", "audio", "hf-asr-leaderboard", "en", "dataset:librispeech_asr", "arxiv:2010.11430", "arxiv:2006.11477", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
"2022-04-12T06:14:30"
--- language: en datasets: - librispeech_asr tags: - speech - audio - automatic-speech-recognition - hf-asr-leaderboard license: apache-2.0 model-index: - name: wav2vec2-large-10min-lv60 results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Librispeech (clean) type: librispeech_asr args: en metrics: - name: Test WER type: wer value: None --- # Wav2Vec2-Large-10min-Lv60 + Self-Training # This is a direct state_dict transfer from fairseq to huggingface, the weights are identical [Facebook's Wav2Vec2](https://ai.facebook.com/blog/wav2vec-20-learning-the-structure-of-speech-from-raw-audio/) The large model pretrained and fine-tuned on 10min of Libri-Light and Librispeech on 16kHz sampled speech audio. Model was trained with [Self-Training objective](https://arxiv.org/abs/2010.11430). When using the model make sure that your speech input is also sampled at 16Khz. [Paper](https://arxiv.org/abs/2006.11477) Authors: Alexei Baevski, Henry Zhou, Abdelrahman Mohamed, Michael Auli **Abstract** They show for the first time that learning powerful representations from speech audio alone followed by fine-tuning on transcribed speech can outperform the best semi-supervised methods while being conceptually simpler. wav2vec 2.0 masks the speech input in the latent space and solves a contrastive task defined over a quantization of the latent representations which are jointly learned. Experiments using all labeled data of Librispeech achieve 1.8/3.3 WER on the clean/other test sets. When lowering the amount of labeled data to one hour, wav2vec 2.0 outperforms the previous state of the art on the 100 hour subset while using 100 times less labeled data. Using just ten minutes of labeled data and pre-training on 53k hours of unlabeled data still achieves 4.8/8.2 WER. This demonstrates the feasibility of speech recognition with limited amounts of labeled data. The original model can be found under https://github.com/pytorch/fairseq/tree/master/examples/wav2vec#wav2vec-20. # Usage To transcribe audio files the model can be used as a standalone acoustic model as follows: ```python from transformers import Wav2Vec2Processor, Wav2Vec2ForCTC from datasets import load_dataset import torch # load model and processor processor = Wav2Vec2Processor.from_pretrained("Splend1dchan/wav2vec2-large-10min-lv60-self") model = Wav2Vec2ForCTC.from_pretrained("Splend1dchan/wav2vec2-large-10min-lv60-self") # load dummy dataset and read soundfiles ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation") # tokenize input_values = processor(ds[0]["audio"]["array"], return_tensors="pt", padding="longest").input_values # retrieve logits logits = model(input_values).logits # take argmax and decode predicted_ids = torch.argmax(logits, dim=-1) transcription = processor.batch_decode(predicted_ids) ``` ## Evaluation This code snippet shows how to evaluate facebook's **Splend1dchan/wav2vec2-large-10min-lv60-self** on LibriSpeech's "clean" and "other" test data. ```python from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import torch from jiwer import wer librispeech_eval = load_dataset("librispeech_asr", "clean", split="test") model = Wav2Vec2ForCTC.from_pretrained("Splend1dchan/wav2vec2-large-10min-lv60-self").to("cuda") processor = Wav2Vec2Processor.from_pretrained("Splend1dchan/wav2vec2-large-10min-lv60-self") def map_to_pred(batch): inputs = processor(batch["audio"]["array"], return_tensors="pt", padding="longest") input_values = inputs.input_values.to("cuda") attention_mask = inputs.attention_mask.to("cuda") with torch.no_grad(): logits = model(input_values, attention_mask=attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) transcription = processor.batch_decode(predicted_ids) batch["transcription"] = transcription return batch result = librispeech_eval.map(map_to_pred, remove_columns=["speech"]) print("WER:", wer(result["text"], result["transcription"])) ``` <!-- *Result (WER)*: | "clean" | "other" | |---|---| | untested | untested | -->
craa/100M_high_500_6910
craa
"2024-12-16T15:16:34"
10
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
"2024-11-01T15:31:54"
--- library_name: transformers tags: - generated_from_trainer metrics: - accuracy model-index: - name: 100M_high_500_6910 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # 100M_high_500_6910 This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.2997 - Accuracy: 0.3947 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0006 - train_batch_size: 32 - eval_batch_size: 16 - seed: 6910 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.98) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:------:|:-----:|:---------------:|:--------:| | 5.1265 | 0.1078 | 1000 | 5.0347 | 0.2260 | | 4.5955 | 0.2156 | 2000 | 4.5171 | 0.2694 | | 4.319 | 0.3235 | 3000 | 4.2408 | 0.2982 | | 4.1618 | 0.4313 | 4000 | 4.0961 | 0.3118 | | 4.0654 | 0.5391 | 5000 | 4.0005 | 0.3204 | | 4.0045 | 0.6469 | 6000 | 3.9213 | 0.3275 | | 3.9296 | 0.7547 | 7000 | 3.8659 | 0.3328 | | 3.8839 | 0.8625 | 8000 | 3.8191 | 0.3370 | | 3.8417 | 0.9704 | 9000 | 3.7820 | 0.3408 | | 3.78 | 1.0782 | 10000 | 3.7566 | 0.3443 | | 3.7404 | 1.1860 | 11000 | 3.7271 | 0.3472 | | 3.7356 | 1.2938 | 12000 | 3.7031 | 0.3493 | | 3.7111 | 1.4016 | 13000 | 3.6790 | 0.3514 | | 3.7009 | 1.5094 | 14000 | 3.6586 | 0.3533 | | 3.6779 | 1.6173 | 15000 | 3.6384 | 0.3554 | | 3.679 | 1.7251 | 16000 | 3.6194 | 0.3569 | | 3.6616 | 1.8329 | 17000 | 3.6055 | 0.3585 | | 3.6429 | 1.9407 | 18000 | 3.5890 | 0.3604 | | 3.5596 | 2.0485 | 19000 | 3.5795 | 0.3617 | | 3.5588 | 2.1563 | 20000 | 3.5717 | 0.3627 | | 3.5599 | 2.2642 | 21000 | 3.5614 | 0.3639 | | 3.5438 | 2.3720 | 22000 | 3.5511 | 0.3646 | | 3.5521 | 2.4798 | 23000 | 3.5406 | 0.3655 | | 3.5404 | 2.5876 | 24000 | 3.5307 | 0.3669 | | 3.5249 | 2.6954 | 25000 | 3.5218 | 0.3676 | | 3.5377 | 2.8032 | 26000 | 3.5126 | 0.3689 | | 3.5209 | 2.9111 | 27000 | 3.5037 | 0.3696 | | 3.4445 | 3.0189 | 28000 | 3.5001 | 0.3707 | | 3.4382 | 3.1267 | 29000 | 3.4963 | 0.3704 | | 3.4678 | 3.2345 | 30000 | 3.4879 | 0.3718 | | 3.4732 | 3.3423 | 31000 | 3.4825 | 0.3725 | | 3.4625 | 3.4501 | 32000 | 3.4759 | 0.3735 | | 3.4536 | 3.5580 | 33000 | 3.4696 | 0.3737 | | 3.4519 | 3.6658 | 34000 | 3.4626 | 0.3748 | | 3.4674 | 3.7736 | 35000 | 3.4575 | 0.3750 | | 3.4363 | 3.8814 | 36000 | 3.4517 | 0.3756 | | 3.4168 | 3.9892 | 37000 | 3.4430 | 0.3762 | | 3.3597 | 4.0970 | 38000 | 3.4475 | 0.3765 | | 3.3832 | 4.2049 | 39000 | 3.4447 | 0.3771 | | 3.3995 | 4.3127 | 40000 | 3.4406 | 0.3771 | | 3.4138 | 4.4205 | 41000 | 3.4348 | 0.3782 | | 3.3852 | 4.5283 | 42000 | 3.4293 | 0.3783 | | 3.4067 | 4.6361 | 43000 | 3.4245 | 0.3792 | | 3.3926 | 4.7439 | 44000 | 3.4207 | 0.3797 | | 3.3885 | 4.8518 | 45000 | 3.4122 | 0.3803 | | 3.3866 | 4.9596 | 46000 | 3.4107 | 0.3806 | | 3.3136 | 5.0674 | 47000 | 3.4116 | 0.3809 | | 3.3054 | 5.1752 | 48000 | 3.4123 | 0.3811 | | 3.3266 | 5.2830 | 49000 | 3.4081 | 0.3816 | | 3.339 | 5.3908 | 50000 | 3.4010 | 0.3818 | | 3.316 | 5.4987 | 51000 | 3.3977 | 0.3822 | | 3.3264 | 5.6065 | 52000 | 3.3936 | 0.3826 | | 3.3376 | 5.7143 | 53000 | 3.3884 | 0.3831 | | 3.3221 | 5.8221 | 54000 | 3.3832 | 0.3837 | | 3.3257 | 5.9299 | 55000 | 3.3806 | 0.3841 | | 3.2366 | 6.0377 | 56000 | 3.3828 | 0.3839 | | 3.2697 | 6.1456 | 57000 | 3.3820 | 0.3843 | | 3.2744 | 6.2534 | 58000 | 3.3811 | 0.3846 | | 3.2915 | 6.3612 | 59000 | 3.3773 | 0.3850 | | 3.28 | 6.4690 | 60000 | 3.3720 | 0.3855 | | 3.2731 | 6.5768 | 61000 | 3.3670 | 0.3859 | | 3.2759 | 6.6846 | 62000 | 3.3628 | 0.3865 | | 3.2859 | 6.7925 | 63000 | 3.3597 | 0.3868 | | 3.2994 | 6.9003 | 64000 | 3.3554 | 0.3870 | | 3.2088 | 7.0081 | 65000 | 3.3595 | 0.3869 | | 3.2185 | 7.1159 | 66000 | 3.3593 | 0.3878 | | 3.2309 | 7.2237 | 67000 | 3.3548 | 0.3876 | | 3.2115 | 7.3315 | 68000 | 3.3523 | 0.3881 | | 3.2466 | 7.4394 | 69000 | 3.3504 | 0.3882 | | 3.2313 | 7.5472 | 70000 | 3.3461 | 0.3887 | | 3.2323 | 7.6550 | 71000 | 3.3410 | 0.3893 | | 3.2287 | 7.7628 | 72000 | 3.3381 | 0.3898 | | 3.2288 | 7.8706 | 73000 | 3.3346 | 0.3899 | | 3.2423 | 7.9784 | 74000 | 3.3312 | 0.3903 | | 3.1552 | 8.0863 | 75000 | 3.3387 | 0.3901 | | 3.1835 | 8.1941 | 76000 | 3.3364 | 0.3901 | | 3.1733 | 8.3019 | 77000 | 3.3324 | 0.3906 | | 3.1908 | 8.4097 | 78000 | 3.3288 | 0.3911 | | 3.1776 | 8.5175 | 79000 | 3.3262 | 0.3913 | | 3.1741 | 8.6253 | 80000 | 3.3220 | 0.3919 | | 3.1867 | 8.7332 | 81000 | 3.3172 | 0.3923 | | 3.1921 | 8.8410 | 82000 | 3.3154 | 0.3926 | | 3.17 | 8.9488 | 83000 | 3.3125 | 0.3928 | | 3.1064 | 9.0566 | 84000 | 3.3167 | 0.3927 | | 3.1415 | 9.1644 | 85000 | 3.3152 | 0.3928 | | 3.1302 | 9.2722 | 86000 | 3.3126 | 0.3933 | | 3.1157 | 9.3801 | 87000 | 3.3108 | 0.3934 | | 3.134 | 9.4879 | 88000 | 3.3070 | 0.3939 | | 3.1284 | 9.5957 | 89000 | 3.3049 | 0.3940 | | 3.131 | 9.7035 | 90000 | 3.3027 | 0.3943 | | 3.1301 | 9.8113 | 91000 | 3.3016 | 0.3945 | | 3.1408 | 9.9191 | 92000 | 3.2997 | 0.3947 | ### Framework versions - Transformers 4.47.0.dev0 - Pytorch 2.5.0+cu124 - Datasets 3.0.2 - Tokenizers 0.20.1
End of preview. Expand in Data Studio

Dataset Card for Hugging Face Hub Model Cards

This datasets consists of model cards for models hosted on the Hugging Face Hub. The model cards are created by the community and provide information about the model, its performance, its intended uses, and more. This dataset is updated on a daily basis and includes publicly available models on the Hugging Face Hub.

This dataset is made available to help support users wanting to work with a large number of Model Cards from the Hub. We hope that this dataset will help support research in the area of Model Cards and their use but the format of this dataset may not be useful for all use cases. If there are other features that you would like to see included in this dataset, please open a new discussion.

Dataset Details

Uses

There are a number of potential uses for this dataset including:

  • text mining to find common themes in model cards
  • analysis of the model card format/content
  • topic modelling of model cards
  • analysis of the model card metadata
  • training language models on model cards

Out-of-Scope Use

[More Information Needed]

Dataset Structure

This dataset has a single split.

Dataset Creation

Curation Rationale

The dataset was created to assist people in working with model cards. In particular it was created to support research in the area of model cards and their use. It is possible to use the Hugging Face Hub API or client library to download model cards and this option may be preferable if you have a very specific use case or require a different format.

Source Data

The source data is README.md files for models hosted on the Hugging Face Hub. We do not include any other supplementary files that may be included in the model card directory.

Data Collection and Processing

The data is downloaded using a CRON job on a daily basis.

Who are the source data producers?

The source data producers are the creators of the model cards on the Hugging Face Hub. This includes a broad variety of people from the community ranging from large companies to individual researchers. We do not gather any information about who created the model card in this repository although this information can be gathered from the Hugging Face Hub API.

Annotations [optional]

There are no additional annotations in this dataset beyond the model card content.

Annotation process

N/A

Who are the annotators?

N/A

Personal and Sensitive Information

We make no effort to anonymize the data. Whilst we don't expect the majority of model cards to contain personal or sensitive information, it is possible that some model cards may contain this information. Model cards may also link to websites or email addresses.

Bias, Risks, and Limitations

Model cards are created by the community and we do not have any control over the content of the model cards. We do not review the content of the model cards and we do not make any claims about the accuracy of the information in the model cards. Some model cards will themselves discuss bias and sometimes this is done by providing examples of bias in either the training data or the responses provided by the model. As a result this dataset may contain examples of bias.

Whilst we do not directly download any images linked to in the model cards, some model cards may include images. Some of these images may not be suitable for all audiences.

Recommendations

Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.

Citation

No formal citation is required for this dataset but if you use this dataset in your work, please include a link to this dataset page.

Dataset Card Authors

@davanstrien

Dataset Card Contact

@davanstrien

Downloads last month
1,392

Collection including librarian-bots/model_cards_with_metadata