▾ G11 Media Network: | ChannelCity | ImpresaCity | SecurityOpenLab | Italian Channel Awards | Italian Project Awards | Italian Security Awards | ...
InnovationOpenLab

Slash AI Energy by 99.5%. rolvsparse© Delivers 20–177× Speedups — Software‑Only, Validated, Patent‑Pending.

rolv, LLC today announced breakthrough benchmarks for rolvsparse©, a patent‑pending software compute primitive delivering 20–177× AI inference speedups and up to 99.5% energy reduction on unmodi...

Immagine

FORT LAUDERDALE, Fla.: rolv, LLC today announced breakthrough benchmarks for rolvsparse©, a patent‑pending software compute primitive delivering 20–177× AI inference speedups and up to 99.5% energy reduction on unmodified models, compared to vendor‑optimized dense and sparse libraries. All energy numbers use real hardware power readings (50 ms polling). No hardware changes, retraining, or precision loss. Most results are independently validated by the University of Miami Frost Institute (some benchmarks are very recent), with bit‑identical SHA‑256 hashes across all platforms.

rolvsparse© achieves 20–177× speedups and 98–99.5% energy savings on real Hugging Face production models. Even fully dense workloads - 0% sparsity - reach 63× acceleration, beating vendor-optimized GPU libraries on CPU alone. Five patents are pending.

“The core idea for rolvsparse came to me on a bike ride in May 2025. Everything since - patents, prototypes, self‑taught stacks, benchmarks, and university validation - has been relentless execution. Why build more data centers when your existing ones can achieve 83× faster performance and 99% greener operations?” - Rolv E. Heggenhougen, Founder & CEO, rolv.ai

Benchmark Highlights (Real Hardware Power Readings), Speedup and Energy savings:

Frontier‑Scale LLMs & MoEs (GPU)

  • Llama‑4 Maverick 400B: 133.5×, 99.9%
  • Llama‑4 400B: 125.3×, 99.4%
  • Llama‑4 400B TTFT: 100.9×
  • Llama‑4 Maverick (MoE): 20.7×, 81.5%
  • Llama‑4 Maverick TTFT: 177×
  • Llama‑4 Scout: 81.7×, 98.8%
  • DeepSeek‑R1 (256 experts): 44.2×, 98.7%
  • Kimi K2.5 (~1T MoE): 10.5×, 90.6%
  • Mixtral 8×22B: 55.1×, 98.2%
  • Claude 3.5‑class FFN: 83×, 98.8%

Qwen Family (GPU + TPU)

  • Qwen3‑235B: 7.8×, 95.5%
  • Qwen2.5‑72B‑Instruct: 50.5×, 91.4%
  • Qwen2.5‑32B (TPU v5e): 5.9×, 83.0%
  • Qwen3.5‑35B GPTQ‑Int4, 64 experts: 9.4×, 89.3%

Specialized & Classical Models

  • GLM‑OCR 0.9B: 50.0×, 98.0%
  • Dense 20k×20k (0% sparse): 63×, 98.42%
  • Finite Element Solver: 163×, 99.5%
  • Netflix Prize RecSys: 3.1×, 67.7%

CPU Benchmarks (Real FFNs)

  • GPT‑J‑6B FFN (40% sparse): 314.6×, 99.7%
  • Mistral‑7B FFN (0% sparse): 253.6×, 99.6%
  • Llama‑2‑7B FFN (70% sparse): 169.2×, 99.4%
  • Kimi K2.5 Expert Slice: 40.3×, 97.9%
  • BERT‑Base FFN (0% sparse): 12.3×, 91.8%

All outputs are SHA‑256 verified for deterministic correctness.

Note: Most CPU benchmarks w/rolv beat GPU w/o rolv.

How rolvsparse© Works

Vendor libraries waste cycles on zeros - the “zero‑FLOP bottleneck.” rolvsparse© restructures arithmetic at the primitive level, skipping meaningless operations while guaranteeing exact outputs (e.g., deterministic hash 8dbe5f139f…dd8dd). It deploys as a drop‑in software layer across all hardware. Users run the open verifier at rolv.ai to generate baselines; rolv returns personalized comparison reports with real power readings.

Market Impact & Applications

AI data centers may reach 9% of U.S. electricity by 2030; hyperscalers have committed $700B+ in AI capex. rolvsparse© reduces energy use by 98–99.5%, boosts throughput on existing infrastructure, and enables edge viability (mobile processors: 70× sparse). Applications include LLMs/MoEs, agents, mobile inference, engineering simulation, RecSys/finance, sustainability, and sovereign AI.

Recent benchmarks on Substack (inl. Json, flops, tokens): rolv.substack.com

Fonte: Business Wire

If you liked this article and want to stay up to date with news from InnovationOpenLab.com subscribe to ours Free newsletter.

Related news

Last News

RSA at Cybertech Europe 2024

Alaa Abdul Nabi, Vice President, Sales International at RSA presents the innovations the vendor brings to Cybertech as part of a passwordless vision for…

Italian Security Awards 2024: G11 Media honours the best of Italian cybersecurity

G11 Media's SecurityOpenLab magazine rewards excellence in cybersecurity: the best vendors based on user votes

How Austria is making its AI ecosystem grow

Always keeping an European perspective, Austria has developed a thriving AI ecosystem that now can attract talents and companies from other countries

Sparkle and Telsy test Quantum Key Distribution in practice

Successfully completing a Proof of Concept implementation in Athens, the two Italian companies prove that QKD can be easily implemented also in pre-existing…

Most read

Roblox Announces Roblox Plus, a New Subscription Plan Offering Exclusive…

Roblox (NYSE: RBLX) today announced Roblox Plus, a new subscription plan designed to deliver more value to its users. Available globally starting April…

Petwealth Emerges from Stealth with $1.7 Million in Funding, Landmark…

#catsofinstagram--Petwealth, the at-home PCR diagnostics and AI-powered health intelligence platform for dogs and cats, today announced its emergence…

From IDP to Intelligent Inference: Hyperscience Hypercell Spring 2026…

Hyperscience, a market leader in enterprise AI infrastructure software, focused on Intelligent Document Processing (IDP), today announced major advancements…

Energy Vault Enters Japanese Market with Acquisition of 850 MW Energy…

$NRGV--Energy Vault Holdings, Inc. (NYSE: NRGV) (“Energy Vault” or the “Company”), a global leader in sustainable, grid-scale energy storage and AI compute…

Newsletter signup

Join our mailing list to get weekly updates delivered to your inbox.

Sign me up!