---license:apache-2.0---## Overview
The Mixtral-7x8B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mistral-7x8Boutperforms Llama 2 70B on most benchmarks we tested.
## Variants
| No | Variant | Cortex CLI command |
| --- | --- | --- |
| 1 | [7x8b-gguf](https://huggingface.co/cortexhub/mixtral/tree/7x8b-gguf) | `cortex run mixtral:7x8b-gguf` |
## Use it with Jan (UI)1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
2. Use in Jan model Hub:
``` cortexhub/mixtral ```## Use it with Cortex (CLI)1. Install **Cortex** using [Quickstart](https://cortex.jan.ai/docs/quickstart)
2. Run the model with command:
``` cortex run mixtral ```## Credits-**Author:** Mistralai
-**Converter:** [Homebrew](https://www.homebrew.ltd/)