mirror of
https://github.com/SakanaAI/doc-to-lora.git
synced 2026-04-25 00:06:20 +02:00
40 lines
1.2 KiB
Bash
40 lines
1.2 KiB
Bash
|
|
WANDB_MODE=disabled run uv run train.py configs/niah_exp/ctx_magic_number_32_256.yaml \
|
||
|
|
--model_name_or_path=Qwen/Qwen3-4B-Instruct-2507 \
|
||
|
|
--num_train_epochs=1 \
|
||
|
|
--per_device_train_batch_size=-1 \
|
||
|
|
--gradient_accumulation_steps=32 \
|
||
|
|
--per_device_eval_batch_size=16 \
|
||
|
|
--exp_setup=hyper_lora \
|
||
|
|
--aggregator_type=perceiver \
|
||
|
|
--target_modules=down_proj \
|
||
|
|
--num_blocks=8 \
|
||
|
|
--num_self_attn_per_block=0 \
|
||
|
|
--num_pre_head_layers=1 \
|
||
|
|
--lora_r=8 \
|
||
|
|
--eval_steps=100 \
|
||
|
|
--logging_steps=10 \
|
||
|
|
--save_steps=1000 \
|
||
|
|
--learning_rate=4e-5 \
|
||
|
|
--lora_dropout=0.0 \
|
||
|
|
--neftune_noise_alpha=0 \
|
||
|
|
--per_rank_gen=True \
|
||
|
|
--per_layer_processing=True \
|
||
|
|
--gen_lora_l1_reg_coef=0.5 \
|
||
|
|
--use_sequence_packing=True \
|
||
|
|
--max_packed_inp_len=2048 \
|
||
|
|
--max_packed_ctx_len=2048 \
|
||
|
|
--dataloader_num_workers=0 \
|
||
|
|
--dataloader_prefetch_factor=None \
|
||
|
|
--eval_on_start=False \
|
||
|
|
--ctx_encoder_type=early_exit \
|
||
|
|
--n_latent_queries=208 \
|
||
|
|
--use_kl_loss=False \
|
||
|
|
--eval_on_start=True \
|
||
|
|
--max_ctx_chunk_len=512 \
|
||
|
|
--min_ctx_chunk_len=25 \
|
||
|
|
--num_chunk_probs='{"1":"0.5", "2":"0.125", "3":"0.0625", "4":"0.0625", "5":"0.0625", "6":"0.0625", "7":"0.0625", "8":"0.0625"}' \
|
||
|
|
--max_val_samples_per_ds=100 \
|
||
|
|
--seed=1 \
|
||
|
|
--use_per_ctx_average_loss=True \
|
||
|
|
--torch_empty_cache_steps=10
|