▾ G11 Media Network: | ChannelCity | ImpresaCity | SecurityOpenLab | Italian Channel Awards | Italian Project Awards | Italian Security Awards | ...
InnovationOpenLab

RunPod Partners with vLLM to Accelerate AI Inference

RunPod, a leading cloud computing platform for AI and machine learning workloads, is excited to announce its partnership with vLLM, a top open-source inference engine. This partnership aims to push th...

Business Wire

Collaboration aims to enhance AI performance and support open-source innovation

MOUNT LAUREL, N.J.: RunPod, a leading cloud computing platform for AI and machine learning workloads, is excited to announce its partnership with vLLM, a top open-source inference engine. This partnership aims to push the boundaries of AI performance and reaffirm RunPod's commitment to the open-source community.

vLLM, known for its innovative PagedAttention algorithm, offers unparalleled efficiency in running large language models. It is widely adopted as the default inference engine for open source large language models across public clouds, model providers, and AI powered products.

As part of this collaboration, RunPod provides compute resources for testing vLLM's inference engine on various GPU models. The partnership also involves regular meetings to discuss AI engineers' needs and ways to advance the field together.

"Our collaboration with vLLM represents a significant step forward in optimizing AI infrastructure," said Zhen Lu, CEO at RunPod. "By supporting vLLM's groundbreaking work, we're not only enhancing AI performance but also reinforcing our dedication to fostering innovation in the open-source community."

The partnership builds on RunPod's involvement with vLLM dating back to summer 2023. This long-term engagement underscores RunPod's commitment to advancing AI technologies and supporting the development of efficient, high-performance tools for AI practitioners.

"vLLM's PagedAttention algorithm is a game-changer in AI inference," added Jean Michael Desrosiers, Head of Customer at RunPod. "It achieves near-optimal memory usage with less than 4% waste, significantly reducing the number of GPUs needed for the same output. This aligns perfectly with our mission to provide efficient, scalable AI infrastructure."

RunPod's support of vLLM extends beyond technical resources. The collaboration aims to create a synergy between RunPod's cloud computing expertise and vLLM's innovative approach to AI inference, potentially leading to new breakthroughs in AI performance and accessibility.

About RunPod:

RunPod is a globally distributed GPU cloud platform that empowers developers to deploy custom full-stack AI applications – simply, globally, and at scale. With RunPod’s key offerings, GPU Instances and Serverless GPUs, developers can develop, train and scale AI applications all in one cloud. RunPod is committed to making cloud computing accessible and affordable without compromising features, usability, or experience. It strives to empower individuals and enterprises with cutting-edge technology, enabling them to unlock the potential of AI and cloud computing. To learn more about RunPod, visit www.runpod.io.

Fonte: Business Wire

If you liked this article and want to stay up to date with news from InnovationOpenLab.com subscribe to ours Free newsletter.

Related news

Last News

RSA at Cybertech Europe 2024

Alaa Abdul Nabi, Vice President, Sales International at RSA presents the innovations the vendor brings to Cybertech as part of a passwordless vision for…

Italian Security Awards 2024: G11 Media honours the best of Italian cybersecurity

G11 Media's SecurityOpenLab magazine rewards excellence in cybersecurity: the best vendors based on user votes

How Austria is making its AI ecosystem grow

Always keeping an European perspective, Austria has developed a thriving AI ecosystem that now can attract talents and companies from other countries

Sparkle and Telsy test Quantum Key Distribution in practice

Successfully completing a Proof of Concept implementation in Athens, the two Italian companies prove that QKD can be easily implemented also in pre-existing…

Most read

Newsletter signup

Join our mailing list to get weekly updates delivered to your inbox.

Sign me up!