| | """ |
| | unit tests for axolotl.core.trainer_builder |
| | """ |
| |
|
| | import pytest |
| |
|
| | from axolotl.core.trainer_builder import HFRLTrainerBuilder |
| | from axolotl.utils.config import normalize_config |
| | from axolotl.utils.dict import DictDefault |
| | from axolotl.utils.models import load_model, load_tokenizer |
| |
|
| |
|
| | @pytest.fixture(name="cfg") |
| | def fixture_cfg(): |
| | cfg = DictDefault( |
| | { |
| | "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6", |
| | "model_type": "AutoModelForCausalLM", |
| | "tokenizer_type": "LlamaTokenizer", |
| | "micro_batch_size": 1, |
| | "gradient_accumulation_steps": 1, |
| | "learning_rate": 0.00005, |
| | "save_steps": 100, |
| | "output_dir": "./model-out", |
| | "warmup_steps": 10, |
| | "gradient_checkpointing": False, |
| | "optimizer": "adamw_torch", |
| | "sequence_len": 2048, |
| | "rl": True, |
| | "adam_beta1": 0.998, |
| | "adam_beta2": 0.9, |
| | "adam_epsilon": 0.00001, |
| | "dataloader_num_workers": 1, |
| | "dataloader_pin_memory": True, |
| | "model_config_type": "llama", |
| | } |
| | ) |
| |
|
| | normalize_config(cfg) |
| |
|
| | return cfg |
| |
|
| |
|
| | @pytest.fixture(name="tokenizer") |
| | def fixture_tokenizer(cfg): |
| | return load_tokenizer(cfg) |
| |
|
| |
|
| | @pytest.fixture(name="model") |
| | def fixture_model(cfg, tokenizer): |
| | return load_model(cfg, tokenizer) |
| |
|
| |
|
| | class TestHFRLTrainerBuilder: |
| | """ |
| | TestCase class for DPO trainer builder |
| | """ |
| |
|
| | def test_build_training_arguments(self, cfg, model, tokenizer): |
| | builder = HFRLTrainerBuilder(cfg, model, tokenizer) |
| | training_arguments = builder.build_training_arguments(100) |
| | assert training_arguments.adam_beta1 == 0.998 |
| | assert training_arguments.adam_beta2 == 0.9 |
| | assert training_arguments.adam_epsilon == 0.00001 |
| | assert training_arguments.dataloader_num_workers == 1 |
| | assert training_arguments.dataloader_pin_memory is True |
| |
|