No description
- Updated training config for Gemma3 1B with CPU offload support - Enhanced progressive_model.py with better error handling - Added support for Mixture-of-Thoughts dataset - Improved compatibility across different server environments 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com> |
||
|---|---|---|
| .devenv | ||
| config | ||
| scripts | ||
| src | ||
| .devenv.flake.nix | ||
| .gitignore | ||
| =2.5.0 | ||
| devenv.lock | ||
| flake-minimal.nix | ||
| flake.lock | ||
| flake.nix | ||
| LORA_TARGET_MODULES.md | ||
| requirements-cpu.txt | ||
| requirements-torch.txt | ||
| requirements.txt | ||
| test_data_load.py | ||