Files changed (1) hide show
  1. README.md +11 -6
README.md CHANGED
@@ -24,22 +24,27 @@ The only difference is that there are a few new training arguments specific to H
24
 
25
  [Here](https://github.com/huggingface/optimum-habana/blob/main/examples/question-answering/run_qa.py) is a question-answering example script to fine-tune a model on SQuAD. You can run it with ALBERT XXL with the following command:
26
  ```bash
27
- python run_qa.py \
28
  --model_name_or_path albert-xxlarge-v1 \
29
- --gaudi_config_name Habana/albert-xxlarge-v1 \
30
  --dataset_name squad \
31
  --do_train \
32
  --do_eval \
33
- --per_device_train_batch_size 12 \
34
  --per_device_eval_batch_size 2 \
35
- --learning_rate 5e-6 \
36
  --num_train_epochs 2 \
37
  --max_seq_length 384 \
38
  --output_dir /tmp/squad/ \
39
  --use_habana \
40
- --use_lazy_mode \
 
 
41
  --throughput_warmup_steps 3 \
42
- --bf16
 
 
 
43
  ```
44
 
45
  Check the [documentation](https://huggingface.co/docs/optimum/habana/index) out for more advanced usage and examples.
 
24
 
25
  [Here](https://github.com/huggingface/optimum-habana/blob/main/examples/question-answering/run_qa.py) is a question-answering example script to fine-tune a model on SQuAD. You can run it with ALBERT XXL with the following command:
26
  ```bash
27
+ PT_HPU_LAZY_MODE=0 PT_ENABLE_INT64_SUPPORT=1 python run_qa.py \
28
  --model_name_or_path albert-xxlarge-v1 \
29
+ --gaudi_config_name Habana/albert-xxlarge-v1
30
  --dataset_name squad \
31
  --do_train \
32
  --do_eval \
33
+ --per_device_train_batch_size 16 \
34
  --per_device_eval_batch_size 2 \
35
+ --learning_rate 7e-05 \
36
  --num_train_epochs 2 \
37
  --max_seq_length 384 \
38
  --output_dir /tmp/squad/ \
39
  --use_habana \
40
+ --use_lazy_mode false \
41
+ --torch_compile \
42
+ --torch_compile_backend hpu_backend \
43
  --throughput_warmup_steps 3 \
44
+ --bf16 \
45
+ --save_strategy no \
46
+ --doc_stride 128 \
47
+ --overwrite_output_di
48
  ```
49
 
50
  Check the [documentation](https://huggingface.co/docs/optimum/habana/index) out for more advanced usage and examples.