diff --git a/img/research/rag-faith.png b/img/research/rag-faith.png new file mode 100644 index 0000000..dbccb25 Binary files /dev/null and b/img/research/rag-faith.png differ diff --git a/pages/models/llama-3.en.mdx b/pages/models/llama-3.en.mdx index f379b4b..3d57665 100644 --- a/pages/models/llama-3.en.mdx +++ b/pages/models/llama-3.en.mdx @@ -1,5 +1,7 @@ # Llama 3 +import {Bleed} from 'nextra-theme-docs' + Meta recently [introduced](https://llama.meta.com/llama3/) their new family of large language models (LLMs) called Llama 3. This release includes 8B and 70B parameters pre-trained and instruction-tuned models. ## Llama 3 Architecture Details @@ -29,4 +31,19 @@ The pretrained models also outperform other models on several benchmarks like AG Meta also reported that they will be releasing a 400B parameter model which is still training and coming soon! There are also efforts around multimodal support, multilingual capabilities, and longer context windows in the pipeline. The current checkpoint for Llama 3 400B (as of April 15, 2024) produces the following results on the common benchmarks like MMLU and Big-Bench Hard: -The licensing information for the Llama 3 models can be found on the [model card](https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md). \ No newline at end of file +!["Llama 3 400B"](../../img/llama3/llama-400b.png) +*Source: [Meta AI](https://ai.meta.com/blog/meta-llama-3/)* + +The licensing information for the Llama 3 models can be found on the [model card](https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md). + +## Extended Review of Llama 3 + +Here is a longer review of Llama 3: + + +