diff --git a/config/training_config_gemma3_1b.yaml b/config/training_config_gemma3_1b.yaml index 2433612..534e5cd 100644 --- a/config/training_config_gemma3_1b.yaml +++ b/config/training_config_gemma3_1b.yaml @@ -91,7 +91,7 @@ progressive_stages: dataset_config: # Mixture-of-Thoughts specific settings streaming: true # Use streaming for large dataset - max_samples: 30000 # Limit samples for faster training + max_samples: 300000 # Limit samples for faster training split: "train" evaluation: diff --git a/flake.nix b/flake.nix index d0d9d05..5bec2f5 100644 --- a/flake.nix +++ b/flake.nix @@ -61,6 +61,12 @@ pytestCheckPhase = "echo 'Skipping tests'"; pythonImportsCheck = []; # Disable import checks }); + curl-cffi = python-super.curl-cffi.overrideAttrs (oldAttrs: { + doCheck = false; + doInstallCheck = false; + pytestCheckPhase = "echo 'Skipping tests'"; + pythonImportsCheck = []; # Disable import checks + }); }; }; }) @@ -192,4 +198,4 @@ LC_ALL = "en_US.UTF-8"; }; }); -} \ No newline at end of file +} diff --git a/src/progressive_model.py b/src/progressive_model.py index b39bd50..99d3876 100644 --- a/src/progressive_model.py +++ b/src/progressive_model.py @@ -381,7 +381,6 @@ class ProgressiveReasoningModel: self.model.save_pretrained(self.adapters[stage_name]) # Also save tokenizer for convenience self.tokenizer.save_pretrained(self.adapters[stage_name]) - def load_for_inference(self, adapter_names: List[str], weights: Optional[Dict[str, float]] = None): """Load model with specific adapters for inference""" if len(adapter_names) == 1: