image/jpeg

Qwen2.5‑Coder‑7B‑Instruct‑Omni1.0 (Isaac Sim Assistant)

Purpose‑built coding assistant for NVIDIA Isaac Sim 5.0+ and Omniverse Kit 107.x. Fine‑tuned to deliver high-level assistance and troubleshooting help.

  • Base: Qwen2.5‑Coder‑7B‑Instruct
  • Repo: TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0
  • Interface: Chat messages or single‑turn text

0. Changelog

  • 2025‑08‑13: Public Inference Endpoint live; added quickstart and examples.

1. Model Introduction

This model specializes in:

  • Isaac Sim API usage and best practices
  • Robot + sensor setup, physics, and extension patterns
  • Robotics code generation and refactoring
  • Diagnosing common Isaac Sim errors and warnings

Key features

  • Chat‑optimized with structure‑aware prompting
  • Good defaults for coding (stable, low randomness)
  • Works via HTTP; no SDK required

2. Model Summary

  • Architecture: Transformer (7B)
  • Context: typical 4K+ (prompt truncation handled internally)
  • Input formats: Chat messages[] or single‑turn inputs
  • Output: generated_text plus simple token accounting

3. Try it now (Public Endpoint)

Current live URL (may change if redeployed):
https://k6yeljf74w9gw134.us-east4.gcp.endpoints.huggingface.cloud

Recommended defaults for coding:

  • temperature: 0.2
  • top_p: 0.7
  • max_new_tokens: 256 (raise as needed)

cURL (chat)

curl -s -X POST "https://k6yeljf74w9gw134.us-east4.gcp.endpoints.huggingface.cloud" \
  -H "Content-Type: application/json" \
  -d '{
    "messages": [
      {"role":"system","content":"You are a helpful coding assistant for NVIDIA Isaac Sim."},
      {"role":"user","content":"Create a minimal script to spawn a URDF and enable PhysX."}
    ],
    "parameters": {"max_new_tokens":256, "temperature":0.2, "top_p":0.9}
  }'

Python

import requests

url = "https://k6yeljf74w9gw134.us-east4.gcp.endpoints.huggingface.cloud"
payload = {
  "messages": [
    {"role":"system","content":"You are a helpful coding assistant for NVIDIA Isaac Sim."},
    {"role":"user","content":"How do I attach a camera sensor to a robot link?"}
  ],
  "parameters": {"max_new_tokens": 256, "temperature": 0.2, "top_p": 0.9}
}
print(requests.post(url, json=payload).json())

Single‑turn (non‑chat)

curl -s -X POST "https://k6yeljf74w9gw134.us-east4.gcp.endpoints.huggingface.cloud" \
  -H "Content-Type: application/json" \
  -d '{"inputs":"Say hello in one sentence.","parameters":{"max_new_tokens":64}}'

4. Inputs / Outputs

Inputs (choose one)

  • Chat: messages as a list of {role, content}
  • Single‑turn: inputs as a string

Common parameters

  • max_new_tokens, temperature, top_p, top_k, repetition_penalty, num_beams, do_sample, seed
  • stop (string or list)
  • max_input_tokens (truncate prompt to reserve room for generation)

Response shape

{
  "generated_text": "...",
  "input_tokens": 123,
  "generated_tokens": 256,
  "total_tokens": 379,
  "params": { "max_new_tokens": 256, "temperature": 0.2 }
}

5. Local usage (Transformers)

from transformers import AutoTokenizer, AutoModelForCausalLM

tok = AutoTokenizer.from_pretrained(
  "TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0", trust_remote_code=True
)
model = AutoModelForCausalLM.from_pretrained(
  "TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0",
  trust_remote_code=True, torch_dtype="auto", device_map="auto"
)

messages = [
  {"role":"system","content":"You are a helpful coding assistant for NVIDIA Isaac Sim."},
  {"role":"user","content":"Example: spawn a robot and start the simulation loop."}
]
prompt = tok.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tok(prompt, return_tensors="pt").to(model.device)
out = model.generate(**inputs, max_new_tokens=256, temperature=0.2)
print(tok.decode(out[0][inputs["input_ids"].shape[-1]:], skip_special_tokens=True))

6. Limitations

  • May produce version‑specific code; verify imports and extension names for your Isaac Sim version.
  • Not a substitute for official safety or hardware guidance.

7. License

MIT (see LICENSE).


8. Citation

@misc{qwen25-coder-isaac-sim,
  title  = {Qwen2.5-Coder-7B-Instruct-Omni1.0: Fine-tuned for NVIDIA Isaac Sim Development},
  author = {TomBombadyl},
  year   = {2024},
  url    = {https://huggingface.co/TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0}
}

9. Contact

Open a Discussion on this model page with questions or feedback.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 5 Ask for provider support

Model tree for TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0

Base model

Qwen/Qwen2.5-7B
Finetuned
(188)
this model

Space using TomBombadyl/Qwen2.5-Coder-7B-Instruct-Omni1.0 1