language:
- en
license: apache-2.0
library_name: transformers
tags:
- merge
- mergekit
- lazymergekit
- mergekit-community/mergekit-della_linear-cwuosuu
- mergekit-community/mergekit-della_linear-nimxtnw
- mergekit-community/mergekit-della_linear-vpjjtsa
- Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
base_model:
- mergekit-community/mergekit-della_linear-cwuosuu
- mergekit-community/mergekit-della_linear-nimxtnw
- mergekit-community/mergekit-della_linear-vpjjtsa
- Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
pipeline_tag: text-generation
model-index:
- name: Llama-3.1-8B-SuperTulu-LexiNova
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: IFEval (0-Shot)
type: HuggingFaceH4/ifeval
args:
num_few_shot: 0
metrics:
- type: inst_level_strict_acc and prompt_level_strict_acc
value: 41.65
name: strict accuracy
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: BBH (3-Shot)
type: BBH
args:
num_few_shot: 3
metrics:
- type: acc_norm
value: 30.5
name: normalized accuracy
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MATH Lvl 5 (4-Shot)
type: hendrycks/competition_math
args:
num_few_shot: 4
metrics:
- type: exact_match
value: 25.3
name: exact match
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GPQA (0-shot)
type: Idavidrein/gpqa
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 4.81
name: acc_norm
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MuSR (0-shot)
type: TAUR-Lab/MuSR
args:
num_few_shot: 0
metrics:
- type: acc_norm
value: 11.23
name: acc_norm
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU-PRO (5-shot)
type: TIGER-Lab/MMLU-Pro
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 26.31
name: accuracy
source:
url: >-
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
name: Open LLM Leaderboard
ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
Overview
ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova is model merge designed to a base for further fine tuning for better natural language understanding and text generation. By combining the best attributes of multiple high-performance models, this fusion allows a highly capable AI with reasoning, compliance, and versatility.
If you want to try the reccomended fine-tuned version of this model, please see here. This model is based on Llama-3.1-8B-Instruct and adheres to the Meta Llama 3.1 Community License Agreement.
π Key Features:
- Enhanced Reasoning & Compliance: Optimized for logical step-by-step thinking.
- Balanced Safety & Utility: Capable of nuanced and detailed responses while maintaining ethical constraints.
- Diverse Knowledge Base: A fusion of models specializing in general instruction, reasoning, and domain-specific tasks.
- Superior Performance: Achieves high benchmarks across multiple evaluations.
π§ Merged Models
This model is a weighted merge of the following:
- Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2 β The foundational model, providing uncensored, high-compliance capabilities.
- mergekit-community/mergekit-della_linear-cwuosuu β Strengthens logical reasoning and alignment.
- mergekit-community/mergekit-della_linear-nimxtnw β Enhances multi-step inference and response depth.
- mergekit-community/mergekit-della_linear-vpjjtsa β Refines contextual understanding and coherence.
π§ Merge Configuration
The following YAML configuration was used to merge these models using Model Stock, ensuring a balanced and optimized fusion:
name: ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
merge_method: model_stock
base_model: Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
dtype: float16
out_dtype: bfloat16
parameters:
normalize: false
int8_mask: true
models:
- model: mergekit-community/mergekit-della_linear-cwuosuu
parameters:
density: 0.5
weight: 0.5
- model: mergekit-community/mergekit-della_linear-nimxtnw
parameters:
density: 0.5
weight: 0.5
- model: mergekit-community/mergekit-della_linear-vpjjtsa
parameters:
density: 0.5
weight: 0.5
- model: Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
parameters:
density: 0.5
weight: 0.5
π How to Use
π₯ Ollama
For quick inference, you can run the model using Ollama:
ollama run hf.co/ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova
π€ Hugging Face Transformers
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
import torch
# Define model name
model_name = "ZeroXClem/Llama-3.1-8B-SuperTulu-LexiNova"
# Load tokenizer & model
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# Initialize text generation pipeline
text_generator = pipeline(
"text-generation",
model=model,
tokenizer=tokenizer,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# Example prompt
prompt = "Explain the importance of AI alignment in modern society."
# Generate output
outputs = text_generator(
prompt,
max_new_tokens=150,
do_sample=True,
temperature=0.7,
top_k=50,
top_p=0.95
)
print(outputs[0]["generated_text"])
π Best Practices
Use System Prompts:
For best results, use a system message before inference:"Think step by step with logical reasoning before providing any response."
For More Uncensored Output:
You can set a different system message or simply use"."
as the system prompt.Quantization Considerations:
Q4
may sometimes cause refusals due to loss in fine-tuning.F16
orQ8
are recommended for optimal performance.
π License
This model is released under the Meta Llama 3.1 Community License Agreement.
Usage, including commercial applications, must adhere to this license.
β Warning: This model is uncensored and highly compliant. Ensure proper alignment layers before deploying as a public service.
π‘ Future Improvements
- Further refinement of reasoning capabilities.
- Optimized token alignment for better coherence.
- Additional quantization tuning for efficient deployment.
β€οΈ Special Thanks
A heartfelt thank you to:
- Orenguteng for Llama-3.1-8B-Lexi-Uncensored-V2.
- MergeKit Community for the powerful della_linear model merges.
- The π€ Hugging Face & Open-Source AI community for advancing AI research.
Your contributions make cutting-edge AI development possible! ππ
π’ Feedback & Contributions
If you encounter any issues or have ideas for improvements, feel free to open a discussion or submit a pull request.
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 23.30 |
IFEval (0-Shot) | 41.65 |
BBH (3-Shot) | 30.50 |
MATH Lvl 5 (4-Shot) | 25.30 |
GPQA (0-shot) | 4.81 |
MuSR (0-shot) | 11.23 |
MMLU-PRO (5-shot) | 26.31 |