@@ -127,32 +127,34 @@ PT_HPU_LAZY_MODE=0 python run_clip.py \
127127Run the following command for distributed training:
128128
129129``` bash
130- PT_HPU_LAZY_MODE=0 \
131- python ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
132- --output_dir ./clip-roberta-finetuned \
133- --model_name_or_path ./clip-roberta \
130+ PT_HPU_LAZY_MODE=0 PT_ENABLE_INT64_SUPPORT=1 \
131+ python3 ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
132+ --output_dir=/tmp/clip_roberta \
133+ --model_name_or_path= ./clip-roberta \
134134 --data_dir $PWD /data \
135135 --dataset_name ydshieh/coco_dataset_script \
136- --dataset_config_name= 2017 \
136+ --dataset_config_name 2017 \
137137 --image_column image_path \
138138 --caption_column caption \
139139 --remove_unused_columns=False \
140- --do_train --do_eval \
141- --per_device_train_batch_size=" 512" \
140+ --do_train --do_eval \
141+ --mediapipe_dataloader \
142+ --per_device_train_batch_size=" 64" \
142143 --per_device_eval_batch_size=" 64" \
143144 --learning_rate=" 5e-5" --warmup_steps=" 0" --weight_decay 0.1 \
144145 --overwrite_output_dir \
145- --save_strategy epoch \
146146 --use_habana \
147- --gaudi_config_name Habana/clip \
148- --throughput_warmup_steps 3 \
149- --dataloader_num_workers 16 \
150- --mediapipe_dataloader \
151- --bf16 \
152- --distribution_strategy fast_ddp \
153- --trust_remote_code \
147+ --use_lazy_mode=False \
148+ --gaudi_config_name= " Habana/clip " \
149+ --throughput_warmup_steps=3 \
150+ --save_strategy= " no " \
151+ --dataloader_num_workers=2 \
152+ --use_hpu_graphs \
153+ --max_steps=100 \
154154 --torch_compile_backend=hpu_backend \
155- --torch_compile
155+ --torch_compile \
156+ --logging_nan_inf_filter \
157+ --trust_remote_code
156158```
157159
158160> ` --mediapipe_dataloader ` only works on Gaudi2.
@@ -163,29 +165,36 @@ python ../gaudi_spawn.py --world_size 8 --use_mpi run_clip.py \
163165Run the following command for training with DeepSpeed:
164166
165167``` bash
166- PT_HPU_LAZY_MODE=0 \
167- python ../gaudi_spawn.py --world_size 8 --use_deepspeed run_clip.py \
168- --output_dir ./clip-roberta-finetuned \
169- --model_name_or_path ./clip-roberta \
168+ PT_HPU_LAZY_MODE=0 PT_ENABLE_INT64_SUPPORT=1 \
169+ python3 ../gaudi_spawn.py --world_size 8 --use_deepspeed run_clip.py \
170+ --output_dir=/tmp/clip_roberta \
171+ --model_name_or_path= ./clip-roberta \
170172 --data_dir $PWD /data \
171173 --dataset_name ydshieh/coco_dataset_script \
172- --dataset_config_name= 2017 \
174+ --dataset_config_name 2017 \
173175 --image_column image_path \
174176 --caption_column caption \
175177 --remove_unused_columns=False \
176- --do_train --do_eval \
177- --per_device_train_batch_size=" 512" \
178+ --do_train --do_eval \
179+ --mediapipe_dataloader \
180+ --per_device_train_batch_size=" 64" \
178181 --per_device_eval_batch_size=" 64" \
179182 --learning_rate=" 5e-5" --warmup_steps=" 0" --weight_decay 0.1 \
180183 --overwrite_output_dir \
181- --save_strategy epoch \
182184 --use_habana \
183- --gaudi_config_name Habana/clip \
184- --throughput_warmup_steps 3 \
185- --deepspeed path_to_my_deepspeed_config \
186- --trust_remote_code \
185+ --use_lazy_mode=False \
186+ --gaudi_config_name=" Habana/clip" \
187+ --throughput_warmup_steps=30 \
188+ --save_strategy=" no" \
189+ --dataloader_num_workers=2 \
190+ --use_hpu_graphs \
191+ --max_steps=100 \
187192 --torch_compile_backend=hpu_backend \
188- --torch_compile
193+ --torch_compile \
194+ --logging_nan_inf_filter \
195+ --trust_remote_code \
196+ --deepspeed < path_to_my_deepspeed_config>
197+
189198```
190199
191200You can look at the [ documentation] ( https://huggingface.co/docs/optimum/habana/usage_guides/deepspeed ) for more information about how to use DeepSpeed in Optimum Habana.
0 commit comments