Skip to content

Commit 39f0a80

Browse files
hsubramonyLiangyx2
authored andcommitted
fix readme cmds for clip-roberta (huggingface#1603)
* fix readme cmds for clip-roberta * comments and cleanup
1 parent b583e3a commit 39f0a80

1 file changed

Lines changed: 38 additions & 29 deletions

File tree

examples/contrastive-image-text/README.md

Lines changed: 38 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -127,32 +127,34 @@ PT_HPU_LAZY_MODE=0 python run_clip.py \
127127
Run the following command for distributed training:
128128

129129
```bash
130-
PT_HPU_LAZY_MODE=0 \
131-
python ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
132-
--output_dir ./clip-roberta-finetuned \
133-
--model_name_or_path ./clip-roberta \
130+
PT_HPU_LAZY_MODE=0 PT_ENABLE_INT64_SUPPORT=1 \
131+
python3 ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
132+
--output_dir=/tmp/clip_roberta \
133+
--model_name_or_path=./clip-roberta \
134134
--data_dir $PWD/data \
135135
--dataset_name ydshieh/coco_dataset_script \
136-
--dataset_config_name=2017 \
136+
--dataset_config_name 2017 \
137137
--image_column image_path \
138138
--caption_column caption \
139139
--remove_unused_columns=False \
140-
--do_train --do_eval \
141-
--per_device_train_batch_size="512" \
140+
--do_train --do_eval \
141+
--mediapipe_dataloader \
142+
--per_device_train_batch_size="64" \
142143
--per_device_eval_batch_size="64" \
143144
--learning_rate="5e-5" --warmup_steps="0" --weight_decay 0.1 \
144145
--overwrite_output_dir \
145-
--save_strategy epoch \
146146
--use_habana \
147-
--gaudi_config_name Habana/clip \
148-
--throughput_warmup_steps 3 \
149-
--dataloader_num_workers 16 \
150-
--mediapipe_dataloader \
151-
--bf16 \
152-
--distribution_strategy fast_ddp \
153-
--trust_remote_code \
147+
--use_lazy_mode=False \
148+
--gaudi_config_name="Habana/clip" \
149+
--throughput_warmup_steps=3 \
150+
--save_strategy="no" \
151+
--dataloader_num_workers=2 \
152+
--use_hpu_graphs \
153+
--max_steps=100 \
154154
--torch_compile_backend=hpu_backend \
155-
--torch_compile
155+
--torch_compile \
156+
--logging_nan_inf_filter \
157+
--trust_remote_code
156158
```
157159

158160
> `--mediapipe_dataloader` only works on Gaudi2.
@@ -163,29 +165,36 @@ python ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
163165
Run the following command for training with DeepSpeed:
164166

165167
```bash
166-
PT_HPU_LAZY_MODE=0 \
167-
python ../gaudi_spawn.py --world_size 8 --use_deepspeed run_clip.py \
168-
--output_dir ./clip-roberta-finetuned \
169-
--model_name_or_path ./clip-roberta \
168+
PT_HPU_LAZY_MODE=0 PT_ENABLE_INT64_SUPPORT=1 \
169+
python3 ../gaudi_spawn.py --world_size 8 --use_deepspeed run_clip.py \
170+
--output_dir=/tmp/clip_roberta \
171+
--model_name_or_path=./clip-roberta \
170172
--data_dir $PWD/data \
171173
--dataset_name ydshieh/coco_dataset_script \
172-
--dataset_config_name=2017 \
174+
--dataset_config_name 2017 \
173175
--image_column image_path \
174176
--caption_column caption \
175177
--remove_unused_columns=False \
176-
--do_train --do_eval \
177-
--per_device_train_batch_size="512" \
178+
--do_train --do_eval \
179+
--mediapipe_dataloader \
180+
--per_device_train_batch_size="64" \
178181
--per_device_eval_batch_size="64" \
179182
--learning_rate="5e-5" --warmup_steps="0" --weight_decay 0.1 \
180183
--overwrite_output_dir \
181-
--save_strategy epoch \
182184
--use_habana \
183-
--gaudi_config_name Habana/clip \
184-
--throughput_warmup_steps 3 \
185-
--deepspeed path_to_my_deepspeed_config \
186-
--trust_remote_code \
185+
--use_lazy_mode=False \
186+
--gaudi_config_name="Habana/clip" \
187+
--throughput_warmup_steps=30 \
188+
--save_strategy="no" \
189+
--dataloader_num_workers=2 \
190+
--use_hpu_graphs \
191+
--max_steps=100 \
187192
--torch_compile_backend=hpu_backend \
188-
--torch_compile
193+
--torch_compile \
194+
--logging_nan_inf_filter \
195+
--trust_remote_code \
196+
--deepspeed <path_to_my_deepspeed_config>
197+
189198
```
190199

191200
You can look at the [documentation](https://huggingface.co/docs/optimum/habana/usage_guides/deepspeed) for more information about how to use DeepSpeed in Optimum Habana.

0 commit comments

Comments
 (0)