[go: up one dir, main page]

DEV Community

# benchmark

Posts

👋 Sign in for the ability to sort posts by relevant, latest, or top.
I benchmarked GPT-4o, Claude 3.5, and Gemini 1.5 for security — the results

I benchmarked GPT-4o, Claude 3.5, and Gemini 1.5 for security — the results

Comments
2 min read
NexusQuant vs KVTC vs TurboQuant vs CommVQ — honest comparison

NexusQuant vs KVTC vs TurboQuant vs CommVQ — honest comparison

Comments
4 min read
I published my benchmark scores. Your turn.

I published my benchmark scores. Your turn.

Comments
4 min read
🚀 8x Faster Than ONNX Runtime: Zero-Allocation AI Inference in Pure C#

🚀 8x Faster Than ONNX Runtime: Zero-Allocation AI Inference in Pure C#

Comments
3 min read
ARC-AGI V3 Explained: The New AI Benchmark That Breaks Every Agent

ARC-AGI V3 Explained: The New AI Benchmark That Breaks Every Agent

Comments
3 min read
GPT-5.1 scored 26%. Gemini 3 Flash scored 74%. Same prompt, same tools.

GPT-5.1 scored 26%. Gemini 3 Flash scored 74%. Same prompt, same tools.

Comments
8 min read
AI Gateways Are Not I/O-Bound Proxies I Benchmarked 5 of Them to Prove It

AI Gateways Are Not I/O-Bound Proxies I Benchmarked 5 of Them to Prove It

2
Comments
9 min read
I Tried Speculative Decoding on RTX 4060 8GB — Every Config Was Slower Than Baseline

I Tried Speculative Decoding on RTX 4060 8GB — Every Config Was Slower Than Baseline

1
Comments
8 min read
FTS vs Hybrid Memory Search: A Real-World Benchmark

FTS vs Hybrid Memory Search: A Real-World Benchmark

1
Comments
4 min read
Token Efficiency: 16 Algorithms, 5 Languages, Zero Guesswork

Token Efficiency: 16 Algorithms, 5 Languages, Zero Guesswork

Comments 1
4 min read
I Built an Auto-Updating Archive of Every AI Arena Leaderboard

I Built an Auto-Updating Archive of Every AI Arena Leaderboard

1
Comments
2 min read
DGX Spark Inference Performance: Local LLM vs Cloud Benchmarks (2026)

DGX Spark Inference Performance: Local LLM vs Cloud Benchmarks (2026)

Comments
5 min read
Running Qwen2.5-32B on RTX 4060 8GB — Beating M4 at 10.8 t/s with llama.cpp

Running Qwen2.5-32B on RTX 4060 8GB — Beating M4 at 10.8 t/s with llama.cpp

1
Comments
7 min read
Benchmarking the Model Is the Wrong Abstraction

Benchmarking the Model Is the Wrong Abstraction

Comments
4 min read
2.78 TFLOPS on a Fanless MacBook Air? Benchmarking Apple's M4 with MLX

2.78 TFLOPS on a Fanless MacBook Air? Benchmarking Apple's M4 with MLX

Comments
4 min read
👋 Sign in for the ability to sort posts by relevant, latest, or top.