top.booster: auto top.checkpoint_path: null top.finetuning_type: full top.model_name: GPT-2-Small top.quantization_bit: none top.quantization_method: bitsandbytes top.rope_scaling: none top.template: default train.additional_target: '' train.apollo_rank: 16 train.apollo_scale: 32 train.apollo_target: all train.apollo_update_interval: 200 train.badam_mode: layer train.badam_switch_interval: 50 train.badam_switch_mode: ascending train.badam_update_ratio: 0.05 train.batch_size: 1 train.compute_type: bf16 train.create_new_adapter: false train.cutoff_len: 1024 train.dataset: - XeTute/Keywords-to-Short-Story - MatanP/emotion_mapped_story_dataset - Chamoda/atlas-storyteller-1000 - jaydenccc/AI_Storyteller_Dataset - webnovel train.dataset_dir: data train.ds_offload: false train.ds_stage: none train.extra_args: '{"optim": "sgd"}' train.freeze_extra_modules: '' train.freeze_trainable_layers: 2 train.freeze_trainable_modules: all train.galore_rank: 16 train.galore_scale: 2 train.galore_target: all train.galore_update_interval: 200 train.gradient_accumulation_steps: 1 train.learning_rate: 1e-1 train.logging_steps: 1000 train.lora_alpha: 16 train.lora_dropout: 0 train.lora_rank: 8 train.lora_target: '' train.loraplus_lr_ratio: 0 train.lr_scheduler_type: cosine train.mask_history: false train.max_grad_norm: '1.0' train.max_samples: '1000000000' train.neat_packing: false train.neftune_alpha: 0 train.num_train_epochs: '4.0' train.packing: false train.ppo_score_norm: false train.ppo_whiten_rewards: false train.pref_beta: 0.1 train.pref_ftx: 0 train.pref_loss: sigmoid train.report_to: - none train.resize_vocab: false train.reward_model: [] train.save_steps: 5000 train.swanlab_api_key: '' train.swanlab_mode: cloud train.swanlab_project: llamafactory train.swanlab_run_name: '' train.swanlab_workspace: '' train.train_on_prompt: false train.training_stage: Supervised Fine-Tuning train.use_apollo: false train.use_badam: false train.use_dora: false train.use_galore: false train.use_llama_pro: false train.use_pissa: false train.use_rslora: false train.use_swanlab: false train.val_size: 0 train.warmup_steps: 0