Latest Announcements
Deepseek OCR: Vision‑Language Compression Meets Dynamic OCR
Published on 2025-11-19More News
Trending Models
All Minilm 33M
All-MiniLM-33M by Sentence Transformers: A compact, monolingual model for efficient information retrieval.
All Minilm 22M
All-MiniLM 22M by Sentence Transformers, a compact model for efficient information retrieval.
Bge M3 567M
Bge M3 567M by BAAI: A bi-lingual, 567M param model for efficient information retrieval with an 8k context window.
Qwen3 0.6B
Qwen3 0.6B: Alibaba's 0.6 billion param model excels in reasoning tasks with a 32k context window.
Deepseek R1 671B
Deepseek R1 671B: Bi-lingual Large Language Model with 671 billion parameters. Supports context lengths up to 128k for advanced code generation and debugging tasks.
Llama3.1 8B Instruct
Llama3.1 8B Instruct by Meta Llama, an 8 billion parameter, multi-lingual model with a 128k context window, excels in assistant-like chat.
Gemma3 4B Instruct
Gemma3 4B by Google: 4 billion params, 128k/32k context-length. Supports multiple languages for creative content generation, chatbot AI, text summarization, and image data extraction.
Qwen3 8B
Qwen3 8B: Alibaba's 8 billion parameter LLM for reasoning & problem-solving. Mono-lingual model with context lengths up to 8k.
Gemma3 1B Instruct
Gemma3 1B by Google: Multilingual LLM with 1B params, 128k/32k context-length. Ideal for content creation & communication tasks.
Llama3 Gradient 8B Instruct
Llama3 Gradient 8B by Meta Llama: An 8 billion parameter, English-focused LLM with an 8k context window, ideal for commercial & research use.
Qwen3 4B
Qwen3 4B: Alibaba's 4 billion parameter LLM for enhanced reasoning & logic tasks.
Qwen3 32B
Qwen3 32B: Alibaba's 32 billion param LLM for complex question answering with logical reasoning.