Louym commited on
Commit
fb043ec
·
verified ·
1 Parent(s): 2ab7bb0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -6,13 +6,12 @@ Here, we provide AWQ-quantized versions of the most popular NVILA models. These
6
  One-command demo to chat with quantized NVILA models via [llm-awq](https://github.com/mit-han-lab/llm-awq/tree/main/tinychat#:~:text=TinyChat%20support%20NVILA) (NVILA-8B as an example):
7
 
8
  ```bash
9
- '''
10
  cd llm-awq/tinychat
11
  python nvila_demo.py --model-path PATH/TO/NVILA \
12
  --quant_path NVILA-8B-w4-g128-awq-v2.pt \
13
  --media PATH/TO/ANY/IMAGES/VIDEOS \
14
  --act_scale_path NVILA-8B-VT-smooth-scale.pt \
15
  --all --chunk --model_type nvila
16
- '''
17
 
18
  This command will download the quantized NVILA model and run a chat demo. If you’ve already downloaded the files, simply set the path to your local copies.
 
6
  One-command demo to chat with quantized NVILA models via [llm-awq](https://github.com/mit-han-lab/llm-awq/tree/main/tinychat#:~:text=TinyChat%20support%20NVILA) (NVILA-8B as an example):
7
 
8
  ```bash
 
9
  cd llm-awq/tinychat
10
  python nvila_demo.py --model-path PATH/TO/NVILA \
11
  --quant_path NVILA-8B-w4-g128-awq-v2.pt \
12
  --media PATH/TO/ANY/IMAGES/VIDEOS \
13
  --act_scale_path NVILA-8B-VT-smooth-scale.pt \
14
  --all --chunk --model_type nvila
15
+ ```
16
 
17
  This command will download the quantized NVILA model and run a chat demo. If you’ve already downloaded the files, simply set the path to your local copies.