WizardLM
WizardLM
AI & ML interests
NLP, LLM
Recent Activity
liked
a model
22 days ago
deepseek-ai/DeepSeek-V3
reacted
to
their
post
with š
6 months ago
š„ š„š„
Excited to announce WizardLM new Paper: Auto Evol-Instruct!
š¦ Twitter: https://x.com/WizardLM_AI/status/1812857977122202087
š Paper: https://arxiv.org/pdf/2406.00770
š¤ 1. Fully AI-Powered Pipeline
Auto Evol-Instruct automatically involves an iterative process of optimizing an Evol-Instruct V1 into an optimal one. The pipeline consists of two critical stages: Evol Trajectory Analysis, where the optimizer LLM analyzes the issues and failures exposed in instruction evolution performed by the evol LLM, and Evolving Method Optimization, where the optimizer LLM addresses these issues to progressively develop an effective evolving method. The optimal evolving method is then used to convert the entire instruction dataset into more diverse and complex forms, facilitating improved instruction tuning.
š2. Scaling Evol-Instruct with Arena Learning
With Auto Evol-Instruct, the evolutionary synthesis data of WizardLM-2 has scaled up from WizardLM-1 to dozens of domains, covering tasks in all aspects of large language models. This allows Arena Learning to train and learn from an almost infinite pool of high-difficulty instruction data, fully unlocking all the potential of Arena Learning.
posted
an
update
6 months ago
š„ š„š„
Excited to announce WizardLM new Paper: Auto Evol-Instruct!
š¦ Twitter: https://x.com/WizardLM_AI/status/1812857977122202087
š Paper: https://arxiv.org/pdf/2406.00770
š¤ 1. Fully AI-Powered Pipeline
Auto Evol-Instruct automatically involves an iterative process of optimizing an Evol-Instruct V1 into an optimal one. The pipeline consists of two critical stages: Evol Trajectory Analysis, where the optimizer LLM analyzes the issues and failures exposed in instruction evolution performed by the evol LLM, and Evolving Method Optimization, where the optimizer LLM addresses these issues to progressively develop an effective evolving method. The optimal evolving method is then used to convert the entire instruction dataset into more diverse and complex forms, facilitating improved instruction tuning.
š2. Scaling Evol-Instruct with Arena Learning
With Auto Evol-Instruct, the evolutionary synthesis data of WizardLM-2 has scaled up from WizardLM-1 to dozens of domains, covering tasks in all aspects of large language models. This allows Arena Learning to train and learn from an almost infinite pool of high-difficulty instruction data, fully unlocking all the potential of Arena Learning.
Organizations
WizardLM's activity
Update README.md
#33 opened about 1 year ago
by
Ziyang
Update README.md
#42 opened about 1 year ago
by
Ziyang
Update README.md
#8 opened about 1 year ago
by
Ziyang
Update README.md
#6 opened about 1 year ago
by
Ziyang
Update README.md
#7 opened about 1 year ago
by
Ziyang
Update README.md
1
#2 opened about 1 year ago
by
ndurkee
Update README.md
#1 opened about 1 year ago
by
Ziyang
Please consider my toiled over coding dataset for fine tuning a 1.1 version of the wizard coder series.
3
#22 opened over 1 year ago
by
rombodawg
Context length is still 4096
3
#7 opened over 1 year ago
by
Shahrukh181
Performance differences with gpt4
#4 opened over 1 year ago
by
Vezora
Phind new model just beat WizardCoder-Python-34B-V1.0 human eval high score
2
#13 opened over 1 year ago
by
rombodawg
Any idea how much VRAM does this use ?
3
#12 opened over 1 year ago
by
Teddydj
Details on the method used to surpass CodeLlama
1
#14 opened over 1 year ago
by
ETN3I
Performance issues when compared to other codellama_34b finetunes
6
#11 opened over 1 year ago
by
rombodawg
wrong bos_token
3
#15 opened over 1 year ago
by
loubnabnl
Build chat bot
3
#16 opened over 1 year ago
by
Hashif
M2 Mac 96gb
1
#17 opened over 1 year ago
by
JordiHugging
How to delete downloaded model from google colab?
#18 opened over 1 year ago
by
kickb
how to install/use
1
#23 opened over 1 year ago
by
IceMasterT
Update config.json
#24 opened over 1 year ago
by
mklf