Latest Announcements

GPT Oss: Configurable Reasoning and Agentic Efficiency
Published on 2025-08-05More News
Trending Models

Gpt Oss 20B
GPT-OSS 20B: OpenAI's 20 billion parameter model for low-latency, local applications. Mono-linguistic.


Gpt Oss 120B
GPT-OSS 120B by OpenAI: 120 billion param. model for complex reasoning tasks, supports English only.


Qwen3 Coder 30B
Qwen3 Coder 30B: Alibaba's 30 billion param LLM for agentic coding, supports mono-lingual contexts up to 2,5,6,k.


Llama3.3 70B Instruct
Llama3.3 70B: Meta's 70 billion param, multi-lingual model for assistant-style chat, with a context window of 128k.


Qwen3 Coder 480B
Qwen3 Coder 480B by Alibaba Qwen: A 480 billion parameter model for monolingual code generation & explanation with context lengths up to 2,5,6,k.


All Minilm 22M
All-MiniLM 22M by Sentence Transformers, a compact model for efficient information retrieval.


All Minilm 33M
All-MiniLM-33M by Sentence Transformers: A compact, monolingual model for efficient information retrieval.


Llama3.2 3B Instruct
Llama3.2 3B Instruct by Meta Llama Enterprise: A multilingual, 3 billion parameter model with 8k to 128k context-length, designed for assistant-like chat apps.


Gemma3 27B Instruct
Gemma3 27B: Google's 27 billion parameter LLM for creative content & comms. Supports context up to 128k tokens. Ideal for text gen, chatbots, summarization & image data extraction.


Qwen2.5Vl 7B
Qwen2.5VL 7B: Alibaba's 7 billion parameter LLM for visual content analysis.


Qwen2.5 Coder 32B Instruct
Qwen2.5 Coder 32B Instruct by Alibaba Qwen: 32 billion param, bilingual, excels in coding tasks.


Mistral Small3.2 24B Instruct
Mistral Small3.2 24B Instruct: A powerful 24-billion parameter model by Mistral AI, designed for chat assistance with a context window of up to 12,800 tokens.
