Latest Announcements
Deepseek OCR: Vision‑Language Compression Meets Dynamic OCR
Published on 2025-11-19More News
Trending Models
Gpt Oss 20B
GPT-OSS 20B: OpenAI's 20 billion parameter model for low-latency, local applications. Mono-linguistic.
Gpt Oss 120B
GPT-OSS 120B by OpenAI: 120 billion param. model for complex reasoning tasks, supports English only.
Qwen3 0.6B
Qwen3 0.6B: Alibaba's 0.6 billion param model excels in reasoning tasks with a 32k context window.
Qwen3 8B
Qwen3 8B: Alibaba's 8 billion parameter LLM for reasoning & problem-solving. Mono-lingual model with context lengths up to 8k.
Qwen3 Coder 30B
Qwen3 Coder 30B: Alibaba's 30 billion param LLM for agentic coding, supports mono-lingual contexts up to 2,5,6,k.
All Minilm 22M
All-MiniLM 22M by Sentence Transformers, a compact model for efficient information retrieval.
All Minilm 33M
All-MiniLM-33M by Sentence Transformers: A compact, monolingual model for efficient information retrieval.
Bge M3 567M
Bge M3 567M by BAAI: A bi-lingual, 567M param model for efficient information retrieval with an 8k context window.
Llama3.1 8B Instruct
Llama3.1 8B Instruct by Meta Llama, an 8 billion parameter, multi-lingual model with a 128k context window, excels in assistant-like chat.
Qwen3 32B
Qwen3 32B: Alibaba's 32 billion param LLM for complex question answering with logical reasoning.
Llama4 109B Instruct
Llama4 109B Instruct by Meta Llama Enterprise: A powerful, multi-lingual large language model with 109 billion parameters, designed for commercial applications.
Qwen3 4B
Qwen3 4B: Alibaba's 4 billion parameter LLM for enhanced reasoning & logic tasks.