dung-vpt-uney commited on
Commit
e942dd7
·
1 Parent(s): b6a01d6

Deploy latest CoRGI Gradio demo

Browse files
Files changed (2) hide show
  1. PROGRESS_LOG.md +1 -0
  2. README.md +36 -7
PROGRESS_LOG.md CHANGED
@@ -8,6 +8,7 @@
8
  - Tidied Qwen generation settings to avoid unused temperature flags when running deterministically.
9
  - Validated ROI extraction on a vision-heavy prompt against the real model and hardened prompts so responses stay in structured JSON without verbose preambles.
10
  - Added meta-comment pruning so thinking-mode rambles (e.g., redundant “Step 3” reflections) are dropped while preserving genuine reasoning; confirmed with the official demo image that only meaningful steps remain.
 
11
 
12
  ## 2024-10-21
13
  - Updated default checkpoints to `Qwen/Qwen3-VL-8B-Thinking` and verified CLI/Gradio/test coverage.
 
8
  - Tidied Qwen generation settings to avoid unused temperature flags when running deterministically.
9
  - Validated ROI extraction on a vision-heavy prompt against the real model and hardened prompts so responses stay in structured JSON without verbose preambles.
10
  - Added meta-comment pruning so thinking-mode rambles (e.g., redundant “Step 3” reflections) are dropped while preserving genuine reasoning; confirmed with the official demo image that only meaningful steps remain.
11
+ - Authored a metadata-rich `README.md` (with Hugging Face Space front matter) so the deployed Space renders without configuration errors.
12
 
13
  ## 2024-10-21
14
  - Updated default checkpoints to `Qwen/Qwen3-VL-8B-Thinking` and verified CLI/Gradio/test coverage.
README.md CHANGED
@@ -1,13 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
  # CoRGI Qwen3-VL Demo
2
 
3
- This Space hosts the CoRGI reasoning pipeline backed by the Qwen/Qwen3-VL-8B-Thinking model.
 
4
 
5
- ## Run Locally
6
- ```
 
 
 
 
 
7
  pip install -r requirements.txt
8
- python examples/demo_qwen_corgi.py
 
 
 
9
  ```
10
 
11
- ## Notes
12
- - The demo queues requests sequentially (ZeroGPU/cpu-basic hardware).
13
- - Configure `CORGI_QWEN_MODEL` to switch to a different checkpoint.
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: CoRGI Qwen3-VL Demo
3
+ emoji: 🐶
4
+ colorFrom: indigo
5
+ colorTo: pink
6
+ sdk: gradio
7
+ sdk_version: "5.41.1"
8
+ app_file: app.py
9
+ pinned: false
10
+ license: apache-2.0
11
+ ---
12
+
13
  # CoRGI Qwen3-VL Demo
14
 
15
+ This Space showcases the CoRGI reasoning pipeline powered entirely by **Qwen/Qwen3-VL-8B-Thinking**.
16
+ Upload an image, ask a visual question, and the app will:
17
 
18
+ 1. Generate structured reasoning steps with visual-verification flags.
19
+ 2. Request region-of-interest evidence for steps that require vision.
20
+ 3. Synthesize a grounded final answer.
21
+
22
+ ## Running Locally
23
+
24
+ ```bash
25
  pip install -r requirements.txt
26
+ python examples/demo_qwen_corgi.py \
27
+ --model-id Qwen/Qwen3-VL-8B-Thinking \
28
+ --max-steps 3 \
29
+ --max-regions 3
30
  ```
31
 
32
+ To launch the Gradio demo locally:
33
+
34
+ ```bash
35
+ python app.py
36
+ ```
37
+
38
+ ## Configuration Notes
39
+
40
+ - The Space queues requests sequentially on `cpu-basic` (ZeroGPU) hardware.
41
+ - Set the `CORGI_QWEN_MODEL` environment variable to try another Qwen3-VL checkpoint (for example, `Qwen/Qwen3-VL-4B-Instruct`).
42
+ - `max_steps` and `max_regions` sliders control how many reasoning steps and ROI candidates the model returns.