[go: up one dir, main page]

1 - 30 of 12,704,291 available results.
model

Google’s latest Gemma, small yet strong for chat and generation

7m

500K+

60

model

Qwen3 is the latest Qwen LLM, built for top-tier coding, math, reasoning, and language tasks.

5m

500K+

152

model

Solid LLaMA 3 update, reliable for coding, chat, and Q&A tasks

11m

100K+

27

model

OpenAI’s open-weight models designed for powerful reasoning, agentic tasks

5m

100K+

43

model

Tiny LLM built for speed, edge devices, and local development

7m

100K+

33

model

Distilled LLaMA by DeepSeek, fast and optimized for real-world tasks

11m

100K+

77

model

Qwen3-Coder is Qwen’s new series of coding agent models.

2m

100K+

25

model

The most advanced Qwen model yet, with major gains in text, vision, video, and reasoning.

5m

100K+

9

model

Versatile Qwen update with better language skills and wider support

12m

100K+

9

model

Google’s latest Gemma, in its QAT (quantization aware trained) variant

7m

100K+

21

model

Efficient multimodal AI for text, image, audio, and video on low-resource devices.

9m

50K+

10

model

Microsoft’s compact model, surprisingly capable at reasoning and code

12m

50K+

22

model

Newest LLama 3 release with improved reasoning and generation quality

11m

50K+

20

model

397B MoE model with 17B activation for reasoning, coding, agents, and multimodal understanding

7d

50K+

5

model

Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use

4m

50K+

2

model

Efficient open model with top-tier performance and fast inference

12m

50K+

20

model

Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.

4m

50K+

1

model

Efficient 80B MoE coding model with 3B activated params, 256K context, and agentic capabilities

2m

10K+

1

model

DeepSeek-V3.2 boosts efficiency and reasoning with DSA, scalable RL, agentic data—IMO/IOI wins.

4m

10K+

10

model

DeepCoder-14B-Preview is a code reasoning LLM fine-tuned to scale up to long context lengths

12m

10K+

13

model

GLM-4.7-Flash is a top 30B-A3B MoE, balancing strong performance with efficient deployment.

2m

10K+

4

model

SmolLM3 is a 3.1B model for efficient on-device use, with strong performance in chat

9m

10K+

7

model

Google’s latest Gemma, small yet strong for chat and generation

5m

10K+

1

model

Meta’s LLama 3.1: Chat-focused, benchmark-strong, multilingual-ready.

1y

10K+

6

model

Qwen3 Embedding: multilingual models for advanced text/ranking tasks like retrieval & clustering.

5m

10K+

model

Image generation model, uses a base latent diffusion model plus a refiner.

2m

10K+

5

model

Ministral 3: compact vision-enabled model with near-24B performance, optimized for local edge use

4m

10K+

4

model

Kimi K2 Thinking: open-source agent with deep reasoning, stable tool use, fast INT4, 256k context.

4m

10K+

1

model

Qwen3 is the latest Qwen LLM, built for top-tier coding, math, reasoning, and language tasks.

5m

10K+

model

Granite Docling is a multimodal model for efficient document conversion.

6m

10K+

2