Post
1891
We recently worked with OpenAI to fine-tune gpt-4o and built the SOTA model for the
patched-codes/static-analysis-eval benchmark. All the code and data
patched-codes/synth-vuln-fixes on how we did it is available on their GitHub - https://github.com/openai/build-hours/tree/main/5-4o_fine_tuning.
Here are some tips based on our experience:
ā Establish baseline with "conditioning" / prompting
ā Task-specific datasets are ideal for PEFT; hard to beat gpt-4o on "broad" tasks
ā Add your best system prompt to each example
ā Ensure training data distribution is similar to inference data
ā Shorten instructions with concise prompts; may require more examples.
ā Define clear evaluation metrics (seriously, please eval!)
You can see more details on the benchmark and process here - https://www.patched.codes/blog/the-static-analysis-evaluation-benchmark-measuring-llm-performance-in-fixing-software-vulnerabilities
Here are some tips based on our experience:
ā Establish baseline with "conditioning" / prompting
ā Task-specific datasets are ideal for PEFT; hard to beat gpt-4o on "broad" tasks
ā Add your best system prompt to each example
ā Ensure training data distribution is similar to inference data
ā Shorten instructions with concise prompts; may require more examples.
ā Define clear evaluation metrics (seriously, please eval!)
You can see more details on the benchmark and process here - https://www.patched.codes/blog/the-static-analysis-evaluation-benchmark-measuring-llm-performance-in-fixing-software-vulnerabilities