o
    	Ti                     @   sJ   d Z ddlmZ ddlmZ ddlmZmZ dd Ze	dkr#e  dS dS )	z
Train Gemma-3 on the Codeforces COTS dataset.

accelerate launch --config_file examples/accelerate_configs/deepspeed_zero3.yaml examples/scripts/sft_gemma3.py
    )load_dataset)AutoModelForImageTextToText)	SFTConfig
SFTTrainerc                  C   st   t ddd} | d} d}tj|dd}t| dd	d	d	d
didddddd
}t||| d}|  |jdd d S )Nzopen-r1/codeforces-cotstrain)splitpromptzgoogle/gemma-3-12b-iteager)attn_implementationz-codeforces-SFTTuse_reentrantFi              )

output_dirbf16use_liger_kernelgradient_checkpointinggradient_checkpointing_kwargs
max_lengthper_device_train_batch_sizegradient_accumulation_stepsdataset_num_procnum_train_epochs)argsmodeltrain_dataset)dataset_name)r   remove_columnsr   from_pretrainedr   r   r   push_to_hub)r   model_idr   training_argstrainer r#   O/home/ubuntu/.local/lib/python3.10/site-packages/examples/scripts/sft_gemma3.pymain!   s.   
r%   __main__N)
__doc__datasetsr   transformersr   trlr   r   r%   __name__r#   r#   r#   r$   <module>   s   !
