Latest Announcements

GPT Oss: Configurable Reasoning and Agentic Efficiency
Published on 2025-08-05More News
Trending Models

Gpt Oss 120B
GPT-OSS 120B by OpenAI: 120 billion param. model for complex reasoning tasks, supports English only.


Qwen3 Coder 30B
Qwen3 Coder 30B: Alibaba's 30 billion param LLM for agentic coding, supports mono-lingual contexts up to 2,5,6,k.


All Minilm 22M
All-MiniLM 22M by Sentence Transformers, a compact model for efficient information retrieval.


All Minilm 33M
All-MiniLM-33M by Sentence Transformers: A compact, monolingual model for efficient information retrieval.


Qwen2.5Vl 7B
Qwen2.5VL 7B: Alibaba's 7 billion parameter LLM for visual content analysis.


Bge M3 567M
Bge M3 567M by BAAI: A bi-lingual, 567M param model for efficient information retrieval with an 8k context window.


Qwen3 0.6B
Qwen3 0.6B: Alibaba's 0.6 billion param model excels in reasoning tasks with a 32k context window.


Llama3.2 1B Instruct
Llama3.2 1B Instruct by Meta Llama Enterprise: A versatile, multilingual large language model with 1 billion parameters and context lengths up to 128k, designed for assistant-like chat applications.


Llama3 8B Instruct
Llama3 8B Instruct by Meta Llama: An 8 billion parameter, English-focused model with an 8k context window, ideal for commercial applications.


Mistral Small3.2 24B Instruct
Mistral Small3.2 24B Instruct: A powerful 24-billion parameter model by Mistral AI, designed for chat assistance with a context window of up to 12,800 tokens.


Gemma3 4B Instruct
Gemma3 4B by Google: 4 billion params, 128k/32k context-length. Supports multiple languages for creative content generation, chatbot AI, text summarization, and image data extraction.


Llama3 8B
Llama3 8B: Meta's 8 billion param, 8k context-length model for English commercial & research use.
