HR-MultiWOZ: A Task Oriented Dialogue (TOD) Dataset for HR LLM Agent Paper • 2402.01018 • Published Feb 1, 2024 • 2
Tulu3 with distraction mitigation data Collection LLM and LRM can be easily distracted by hidden instructions or irrelevant tasks. We curated SFT and DPO data that model can finetune to avoid distract • 5 items • Updated 21 days ago • 2
groupfairnessllm/tulu-3-sft-personas-code-with-distraction Viewer • Updated about 1 month ago • 1.7k • 24
groupfairnessllm/tulu-3-sft-personas-instruction-following-with-distraction Viewer • Updated about 1 month ago • 1.7k • 34
groupfairnessllm/tulu-3-sft-personas-math-with-distraction Viewer • Updated about 1 month ago • 1.7k • 22
groupfairnessllm/tulu-3-preference-personas-math-with-distraction Viewer • Updated about 1 month ago • 500 • 32
groupfairnessllm/tulu-3-preference-personas-instruction-following-with-distraction Viewer • Updated about 1 month ago • 500 • 22
Tulu3 with distraction mitigation data Collection LLM and LRM can be easily distracted by hidden instructions or irrelevant tasks. We curated SFT and DPO data that model can finetune to avoid distract • 5 items • Updated 21 days ago • 2
Distractor Injection Attacks on Large Reasoning Models: Characterization and Defense Paper • 2510.16259 • Published Oct 17 • 3
Tulu3 with distraction mitigation data Collection LLM and LRM can be easily distracted by hidden instructions or irrelevant tasks. We curated SFT and DPO data that model can finetune to avoid distract • 5 items • Updated 21 days ago • 2
Tulu3 with distraction mitigation data Collection LLM and LRM can be easily distracted by hidden instructions or irrelevant tasks. We curated SFT and DPO data that model can finetune to avoid distract • 5 items • Updated 21 days ago • 2
groupfairnessllm/tulu-3-preference-personas-math-with-distraction Viewer • Updated about 1 month ago • 500 • 32
groupfairnessllm/tulu-3-preference-personas-instruction-following-with-distraction Viewer • Updated about 1 month ago • 500 • 22