--- license: apache-2.0 language: - en base_model: - Qwen/Qwen2.5-Math-1.5B pipeline_tag: question-answering library_name: transformers tags: - verifier --- This is the verifier we used in [General Reasoner](https://github.com/TIGER-AI-Lab/General-Reasoner). ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM import torch # Replace with your model path model_path = "TIGER-Lab/general-verifier" # Load tokenizer and model tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=torch.float16).cuda() # Example inputs question = "Factor the following quadratic: $3 x^3+\frac{69 x^2}{2}-36 x-810$" ground_truth = "\\frac{3(2x-9)(x+6)(x+10)}{2}" student_answer = "\\frac{3}{2}(x+6)(2x-9)(x+10)" # Create prompt prompt = ( f"User: ### Question: {question}\n\n" f"### Ground Truth Answer: {ground_truth}\n\n" f"### Student Answer: {student_answer}\n\n" "For the above question, please verify if the student's answer is equivalent to the ground truth answer.\n" "Do not solve the question by yourself; just check if the student's answer is equivalent to the ground truth answer.\n" "If the student's answer is correct, output \"Final Decision: Yes\". If the student's answer is incorrect, output \"Final Decision: No\". Assistant:" ) # Tokenize and generate inputs = tokenizer(prompt, return_tensors="pt").to(model.device) outputs = model.generate( **inputs, max_new_tokens=1024, temperature=0.0, do_sample=False ) # Decode and print output print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ```