Joblib
English
llm
human-feedback
weak supervision
data filtering
cmglaze commited on
Commit
279205a
·
1 Parent(s): ae10b2a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -2
README.md CHANGED
@@ -24,7 +24,7 @@ tags:
24
  # Summary
25
  Instruction tuning has emerged as an important step in developing performant large language models (LLMs) for generative AI tasks. While industry-backed LLMs such as ChatGPT, Bard, Claude, and even the open-source Llama 2 have relied on massive, expensive proprietary datasets unavailable to the public, the open source community has banded together to create similar datasets such as OpenAssistant and Dolly that are available to everyone. However, high variance in the quality and distribution of responses collected by volunteers has limited the quality of resulting open source models.
26
 
27
- This model (1) classifies instruction with a standardized schema that can be applied across datasets and (2) scores response quality on a scale of 0-1. The purpose is to measure and track instruction diversity across training sets, and enable filtering based on response quality for more targeted fine-tuning.
28
 
29
  The instruction classification schema is based on prior work in large language models:
30
 
@@ -35,8 +35,15 @@ The instruction classification schema is based on prior work in large language m
35
  * <strong>Summarization</strong>: e.g., “Summarize the main points from this news article”
36
  * <strong>Other</strong>: e.g., anything that did not fit the previous five categories.
37
 
 
 
 
 
 
 
 
38
  # Model development
39
- The model consists of a chain of two xbgoost algorithms, one for instruction classification and one for response quality (modeled as binary is/is not an acceptable response classifier). We trained the algorithms with weak supervision techniques and a feature space that includes metadata specific to each dataset, perplexity, measures with [simcse embeddings](https://arxiv.org/pdf/2104.08821.pdf) and attributes involving regex, parts-of-speech tagging and response duration. In order to maintain a lightweight architecture that requires only CPUs for inference, we omitted perplexity from the feature space used to train the xgboost end models (so perplexity was used for weak supervision only).
40
 
41
  # Model evaluation
42
  ## Instruction classification
 
24
  # Summary
25
  Instruction tuning has emerged as an important step in developing performant large language models (LLMs) for generative AI tasks. While industry-backed LLMs such as ChatGPT, Bard, Claude, and even the open-source Llama 2 have relied on massive, expensive proprietary datasets unavailable to the public, the open source community has banded together to create similar datasets such as OpenAssistant and Dolly that are available to everyone. However, high variance in the quality and distribution of responses collected by volunteers has limited the quality of resulting open source models.
26
 
27
+ This model (1) classifies instructions with a standardized schema that can be applied across datasets and (2) scores response quality on a scale of 0-1. The purpose is to measure and track instruction diversity across training sets, and enable filtering based on response quality for more targeted fine-tuning.
28
 
29
  The instruction classification schema is based on prior work in large language models:
30
 
 
35
  * <strong>Summarization</strong>: e.g., “Summarize the main points from this news article”
36
  * <strong>Other</strong>: e.g., anything that did not fit the previous five categories.
37
 
38
+ The response quality model was developed as a binary classlifier ("is/is not an acceptable response"), with the following goals:
39
+ 1. Enable filtering of instruction/response datasets for higher quality responses.
40
+ 2. Enable this filtering while maintaining (or increasing) instruction diversity.
41
+ 3. Develop the model with a lightweight architecture and a scalable data labeling process that requires minimal human-hours.
42
+
43
+ The response model itself was developed with weak supervision in one day by two FTEs at Snorkel, AI. The model is under continuous development and planned work includes additional datasets and more refined curating labeling functions. We also welcome feedback from the community on any observed patterns of errors!
44
+
45
  # Model development
46
+ The model pipeline currently consists of a chain of two xbgoost algorithms, one for instruction classification and one for response quality (modeled as binary is/is not an acceptable response classifier). We trained the algorithms with weak supervision techniques and a feature space that includes metadata specific to each dataset, perplexity, measures with [simcse embeddings](https://arxiv.org/pdf/2104.08821.pdf) and attributes involving regex, parts-of-speech tagging and response duration. In order to maintain a lightweight architecture that requires only CPUs for inference, we omitted perplexity from the feature space used to train the xgboost end models (so perplexity was used for weak supervision only).
47
 
48
  # Model evaluation
49
  ## Instruction classification