▾ G11 Media Network: | ChannelCity | ImpresaCity | SecurityOpenLab | Italian Channel Awards | Italian Project Awards | Italian Security Awards | ...
InnovationOpenLab

Slash AI Energy by 99.5%. rolvsparse© Delivers 20–177× Speedups — Software‑Only, Validated, Patent‑Pending.

rolv, LLC today announced breakthrough benchmarks for rolvsparse©, a patent‑pending software compute primitive delivering 20–177× AI inference speedups and up to 99.5% energy reduction on unmodi...

Immagine

FORT LAUDERDALE, Fla.: rolv, LLC today announced breakthrough benchmarks for rolvsparse©, a patent‑pending software compute primitive delivering 20–177× AI inference speedups and up to 99.5% energy reduction on unmodified models, compared to vendor‑optimized dense and sparse libraries. All energy numbers use real hardware power readings (50 ms polling). No hardware changes, retraining, or precision loss. Most results are independently validated by the University of Miami Frost Institute (some benchmarks are very recent), with bit‑identical SHA‑256 hashes across all platforms.

rolvsparse© achieves 20–177× speedups and 98–99.5% energy savings on real Hugging Face production models. Even fully dense workloads - 0% sparsity - reach 63× acceleration, beating vendor-optimized GPU libraries on CPU alone. Five patents are pending.

“The core idea for rolvsparse came to me on a bike ride in May 2025. Everything since - patents, prototypes, self‑taught stacks, benchmarks, and university validation - has been relentless execution. Why build more data centers when your existing ones can achieve 83× faster performance and 99% greener operations?” - Rolv E. Heggenhougen, Founder & CEO, rolv.ai

Benchmark Highlights (Real Hardware Power Readings), Speedup and Energy savings:

Frontier‑Scale LLMs & MoEs (GPU)

  • Llama‑4 Maverick 400B: 133.5×, 99.9%
  • Llama‑4 400B: 125.3×, 99.4%
  • Llama‑4 400B TTFT: 100.9×
  • Llama‑4 Maverick (MoE): 20.7×, 81.5%
  • Llama‑4 Maverick TTFT: 177×
  • Llama‑4 Scout: 81.7×, 98.8%
  • DeepSeek‑R1 (256 experts): 44.2×, 98.7%
  • Kimi K2.5 (~1T MoE): 10.5×, 90.6%
  • Mixtral 8×22B: 55.1×, 98.2%
  • Claude 3.5‑class FFN: 83×, 98.8%

Qwen Family (GPU + TPU)

  • Qwen3‑235B: 7.8×, 95.5%
  • Qwen2.5‑72B‑Instruct: 50.5×, 91.4%
  • Qwen2.5‑32B (TPU v5e): 5.9×, 83.0%
  • Qwen3.5‑35B GPTQ‑Int4, 64 experts: 9.4×, 89.3%

Specialized & Classical Models

  • GLM‑OCR 0.9B: 50.0×, 98.0%
  • Dense 20k×20k (0% sparse): 63×, 98.42%
  • Finite Element Solver: 163×, 99.5%
  • Netflix Prize RecSys: 3.1×, 67.7%

CPU Benchmarks (Real FFNs)

  • GPT‑J‑6B FFN (40% sparse): 314.6×, 99.7%
  • Mistral‑7B FFN (0% sparse): 253.6×, 99.6%
  • Llama‑2‑7B FFN (70% sparse): 169.2×, 99.4%
  • Kimi K2.5 Expert Slice: 40.3×, 97.9%
  • BERT‑Base FFN (0% sparse): 12.3×, 91.8%

All outputs are SHA‑256 verified for deterministic correctness.

Note: Most CPU benchmarks w/rolv beat GPU w/o rolv.

How rolvsparse© Works

Vendor libraries waste cycles on zeros - the “zero‑FLOP bottleneck.” rolvsparse© restructures arithmetic at the primitive level, skipping meaningless operations while guaranteeing exact outputs (e.g., deterministic hash 8dbe5f139f…dd8dd). It deploys as a drop‑in software layer across all hardware. Users run the open verifier at rolv.ai to generate baselines; rolv returns personalized comparison reports with real power readings.

Market Impact & Applications

AI data centers may reach 9% of U.S. electricity by 2030; hyperscalers have committed $700B+ in AI capex. rolvsparse© reduces energy use by 98–99.5%, boosts throughput on existing infrastructure, and enables edge viability (mobile processors: 70× sparse). Applications include LLMs/MoEs, agents, mobile inference, engineering simulation, RecSys/finance, sustainability, and sovereign AI.

Recent benchmarks on Substack (inl. Json, flops, tokens): rolv.substack.com

Fonte: Business Wire

If you liked this article and want to stay up to date with news from InnovationOpenLab.com subscribe to ours Free newsletter.

Related news

Last News

RSA at Cybertech Europe 2024

Alaa Abdul Nabi, Vice President, Sales International at RSA presents the innovations the vendor brings to Cybertech as part of a passwordless vision for…

Italian Security Awards 2024: G11 Media honours the best of Italian cybersecurity

G11 Media's SecurityOpenLab magazine rewards excellence in cybersecurity: the best vendors based on user votes

How Austria is making its AI ecosystem grow

Always keeping an European perspective, Austria has developed a thriving AI ecosystem that now can attract talents and companies from other countries

Sparkle and Telsy test Quantum Key Distribution in practice

Successfully completing a Proof of Concept implementation in Athens, the two Italian companies prove that QKD can be easily implemented also in pre-existing…

Most read

Genspark Claw Launches as Genspark’s First “AI Employee,” Alongside Genspark…

Genspark.ai today announced the launch of Genspark Claw, introduced as users’ first “AI employee.” Genspark Claw allows users to delegate work via a simple…

Juicebox Raises $80M at $850M Valuation to Help Businesses Reach Top Talent…

Juicebox, the AI recruiting platform, today announced $80 million in Series B funding at an $850 million valuation led by DST Global, with meaningful…

United Rentals Introduces AI-Powered Equipment Agent

United Rentals, Inc. (NYSE: URI) today announced the launch of the Equipment Agent, a first-of-its-kind AI-powered equipment recommendation solution designed…

OpenFold Consortium Announces Major OpenFold3 Update and Public Release…

The OpenFold Consortium today announced a major OpenFold3 update and the public release of training datasets and full-stack tooling for reproducible biomolecular…

Newsletter signup

Join our mailing list to get weekly updates delivered to your inbox.

Sign me up!