No description
| .devenv | ||
| config | ||
| docs | ||
| scripts | ||
| src | ||
| .devenv.flake.nix | ||
| .env.example | ||
| .gitignore | ||
| .python-version | ||
| =2.5.0 | ||
| devenv.lock | ||
| flake-minimal.nix | ||
| flake.lock | ||
| flake.nix | ||
| LORA_TARGET_MODULES.md | ||
| pyproject.toml | ||
| README.md | ||
| requirements-cpu.txt | ||
| requirements-torch.txt | ||
| requirements.txt | ||
| test_data_load.py | ||
| uv.lock | ||
Progressive LLM Training
Progressive training for LLMs with 8-GPU support for 松尾研LLMコンペ2025.
Quick Start
# Install uv
curl -LsSf https://astral.sh/uv/install.sh | sh
# Setup project
git clone <repository-url>
cd progressive-llm-training
uv sync
# Start training
uv run scripts/train_progressive.py --config config/training_config_gemma3_1b.yaml
./scripts/train_gemma3_1b_8gpu.sh --strategy deepspeed
Training Stages
- basic_cot - Basic reasoning
- math_reasoning - Math with OpenR1-Math-220k
- complex_reasoning - Complex reasoning with Mixture-of-Thoughts
Commands
uv sync # Install dependencies
uv run scripts/train_progressive.py --config config/training_config_gemma3_1b.yaml # Single GPU
./scripts/train_gemma3_1b_8gpu.sh --strategy deepspeed # 8 GPUs
uv run pytest # Run tests
Key Files
config/training_config_gemma3_1b_8gpu_deepspeed.yaml- 8-GPU configscripts/train_progressive.py- Main training scriptscripts/train_gemma3_1b_8gpu.sh- 8-GPU launchersrc/progressive_model.py- Core model implementation
Ready to train! 🚀