Apache-2.0 Open High Quality Math Corpus
AI & ML interests
None defined yet.
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code
-
tokyotech-llm/swallow-code
Viewer • Updated • 129M • 11k • 59 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0002500
Updated • 3 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0005000
8B • Updated • 2 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0007500
8B • Updated • 6
-
tokyotech-llm/Llama-3-Swallow-8B-v0.1
Text Generation • Updated • 284 • • 12 -
tokyotech-llm/Llama-3-Swallow-70B-v0.1
Text Generation • Updated • 12 • • 6 -
tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1
Text Generation • 8B • Updated • 15.9k • • 21 -
tokyotech-llm/Llama-3-Swallow-70B-Instruct-v0.1
Text Generation • 71B • Updated • 17 • • 7
Swallow instruction tuning models
-
tokyotech-llm/Swallow-7b-instruct-hf
Text Generation • 7B • Updated • 4.22k • 44 -
tokyotech-llm/Swallow-13b-instruct-v0.1
Text Generation • 13B • Updated • 57 • 1 -
tokyotech-llm/Swallow-70b-instruct-v0.1
Text Generation • 69B • Updated • 3 -
tokyotech-llm/Swallow-7b-instruct-v0.1
Text Generation • 7B • Updated • 107 • 3
Swallow MX(Mixtral) models
Apache-2.0 Open High Quality Code Corpus
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code
-
tokyotech-llm/swallow-math
Viewer • Updated • 4.33M • 1.02k • 38 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0002500
8B • Updated • 3 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0005000
8B • Updated • 1 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0007500
8B • Updated
-
tokyotech-llm/Gemma-2-Llama-Swallow-27b-pt-v0.1
Text Generation • 27B • Updated • 108 • 1 -
tokyotech-llm/Gemma-2-Llama-Swallow-9b-pt-v0.1
Text Generation • Updated • 2.1k • 1 -
tokyotech-llm/Gemma-2-Llama-Swallow-2b-pt-v0.1
Text Generation • Updated • 67 -
tokyotech-llm/Gemma-2-Llama-Swallow-2b-it-v0.1
Text Generation • Updated • 142 • 4
-
tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5
Text Generation • 8B • Updated • 12.9k • • 18 -
tokyotech-llm/Llama-3.1-Swallow-8B-v0.5
8B • Updated • 4.02k • 9 -
tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
Text Generation • 71B • Updated • 419 • 14 -
tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3
Text Generation • 8B • Updated • 3.46k • • 24
Continual Pre-Training from Llama 2
Swallow MS/MX (Mistral/Mixtral) models
Apache-2.0 Open High Quality Math Corpus
Apache-2.0 Open High Quality Code Corpus
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code
-
tokyotech-llm/swallow-math
Viewer • Updated • 4.33M • 1.02k • 38 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0002500
8B • Updated • 3 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0005000
8B • Updated • 1 -
tokyotech-llm/Llama-3.1-8B-math-ablation-exp2-LR2.5e-5-WD0.1-iter0007500
8B • Updated
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code
-
tokyotech-llm/swallow-code
Viewer • Updated • 129M • 11k • 59 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0002500
Updated • 3 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0005000
8B • Updated • 2 -
tokyotech-llm/Llama-3.1-8B-code-ablation-exp1-LR2.5e-5-MINLR2.5E-6-WD0.1-iter0007500
8B • Updated • 6
-
tokyotech-llm/Gemma-2-Llama-Swallow-27b-pt-v0.1
Text Generation • 27B • Updated • 108 • 1 -
tokyotech-llm/Gemma-2-Llama-Swallow-9b-pt-v0.1
Text Generation • Updated • 2.1k • 1 -
tokyotech-llm/Gemma-2-Llama-Swallow-2b-pt-v0.1
Text Generation • Updated • 67 -
tokyotech-llm/Gemma-2-Llama-Swallow-2b-it-v0.1
Text Generation • Updated • 142 • 4
-
tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5
Text Generation • 8B • Updated • 12.9k • • 18 -
tokyotech-llm/Llama-3.1-Swallow-8B-v0.5
8B • Updated • 4.02k • 9 -
tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
Text Generation • 71B • Updated • 419 • 14 -
tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3
Text Generation • 8B • Updated • 3.46k • • 24
-
tokyotech-llm/Llama-3-Swallow-8B-v0.1
Text Generation • Updated • 284 • • 12 -
tokyotech-llm/Llama-3-Swallow-70B-v0.1
Text Generation • Updated • 12 • • 6 -
tokyotech-llm/Llama-3-Swallow-8B-Instruct-v0.1
Text Generation • 8B • Updated • 15.9k • • 21 -
tokyotech-llm/Llama-3-Swallow-70B-Instruct-v0.1
Text Generation • 71B • Updated • 17 • • 7
Continual Pre-Training from Llama 2
Swallow instruction tuning models
-
tokyotech-llm/Swallow-7b-instruct-hf
Text Generation • 7B • Updated • 4.22k • 44 -
tokyotech-llm/Swallow-13b-instruct-v0.1
Text Generation • 13B • Updated • 57 • 1 -
tokyotech-llm/Swallow-70b-instruct-v0.1
Text Generation • 69B • Updated • 3 -
tokyotech-llm/Swallow-7b-instruct-v0.1
Text Generation • 7B • Updated • 107 • 3
Swallow MS/MX (Mistral/Mixtral) models
Swallow MX(Mixtral) models