Sort By Popular Newest
-
Phi-4 is a 14B parameter, state-of-the-art open model from Microsoft.
-
Llama 3.1 is a new state-of-the-art model from Meta available in 8B, 70B and 405B parameter sizes.
-
A high-performing open embedding model with a large token context window.
-
The 7B model released by Mistral AI, updated to version 0.3.
-
Qwen 1.5 is a series of large language models by Alibaba Cloud spanning from 0.5B to 110B parameters
-
Qwen2 is a new series of large language models from Alibaba group
-
Google Gemma 2 is a high-performing and efficient model available in three sizes: 2B, 9B, and 27B.
-
Llama 2 is a collection of foundation language models ranging from 7B to 70B parameters.
-
A large language model that can use text prompts to generate and discuss code.
-
The TinyLlama project is an open endeavor to train a compact 1.1B Llama model on 3 trillion tokens.
-
A suite of text embedding models by Snowflake, optimized for performance.
-
DeepSeek Coder is a capable coding model trained on two trillion code and natural language tokens.