
Granite-3.1-8B-Instruct is a 8B parameter long-context instruct model finetuned from Granite-3.1-8B-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets tailored for solving long context problems. This model is developed using a diverse set of techniques with a structured chat format, including supervised finetuning, model alignment using reinforcement learning, and model merging.
- Developers: Granite Team, IBM
- GitHub Repository: ibm-granite/granite-3.1-language-models
- Website: Granite Docs
- Paper: Granite 3.1 Language Models (coming soon)
- Release Date: December 18th, 2024
- License: Apache 2.0
Supported Languages: English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, and Chinese. Users may finetune Granite 3.1 models for languages beyond these 12 languages.
Note: This table showed source model instead of quantized model evaluation. Source Model Evaluation refer to Granite-3.1-8b-Instruct Evaluation Result
Models | ARC-Challenge | Hellaswag | MMLU | TruthfulQA | Winogrande | GSM8K | Avg |
---|---|---|---|---|---|---|---|
Granite-3.1-8B-Instruct | 62.62 | 84.48 | 65.34 | 66.23 | 75.37 | 73.84 | 71.31 |
Granite-3.1-2B-Instruct | 54.61 | 75.14 | 55.31 | 59.42 | 67.48 | 52.76 | 60.79 |
Granite-3.1-3B-A800M-Instruct | 50.42 | 73.01 | 52.19 | 49.71 | 64.87 | 48.97 | 56.53 |
Granite-3.1-1B-A400M-Instruct | 42.66 | 65.97 | 26.13 | 46.77 | 62.35 | 33.88 | 46.29 |
Models | IFEval | BBH | MATH Lvl 5 | GPQA | MUSR | MMLU-Pro | Avg |
---|---|---|---|---|---|---|---|
Granite-3.1-8B-Instruct | 72.08 | 34.09 | 21.68 | 8.28 | 19.01 | 28.19 | 30.55 |
Granite-3.1-2B-Instruct | 62.86 | 21.82 | 11.33 | 5.26 | 4.87 | 20.21 | 21.06 |
Granite-3.1-3B-A800M-Instruct | 55.16 | 16.69 | 10.35 | 5.15 | 2.51 | 12.75 | 17.1 |
Granite-3.1-1B-A400M-Instruct | 46.86 | 6.18 | 4.08 | 0 | 0.78 | 2.41 | 10.05 |
Users can run large language models on Qualcomm chips using either of the following methods:
Run large models with APLUX AidGen: Please refer to the APLUX AidGen Developer Documentation
Run large models with Qualcomm Genie: Please refer to the Qualcomm Genie Documentation