Skip to Content Facebook Feature Image

Penguin Solutions’ OriginAI Factory Platform Delivers Optimized Performance for AI Inference

News

Penguin Solutions’ OriginAI Factory Platform Delivers Optimized Performance for AI Inference
News

News

Penguin Solutions’ OriginAI Factory Platform Delivers Optimized Performance for AI Inference

2026-03-17 04:38 Last Updated At:04:51

FREMONT, Calif.--(BUSINESS WIRE)--Mar 16, 2026--

Penguin Solutions, Inc. (Nasdaq: PENG ), the AI factory platform company, today announced the expansion of its OriginAI ® portfolio to include solutions that address the need for more GPU memory to solve context size and concurrency, and meet low latency demands of enterprise-scale AI inference. Penguin Solutions’ OriginAI inference solutions seamlessly add large memory appliances to NVIDIA RTX PRO 6000 and NVIDIA B300 GPU designs, helping to shorten time to value and mitigate performance limitations of AI inference. Designed to improve key operational metrics such as GPU utilization, deployment velocity, and infrastructure reliability, OriginAI enables organizations to run AI workloads with predictable performance at scale.

This press release features multimedia. View the full release here: https://www.businesswire.com/news/home/20260316410520/en/

OriginAI inference solutions are designed leveraging Penguin Solutions 3.3+ billion hours of GPU runtime experience and more than 30 years of expertise delivering advanced memory solutions. OriginAI delivers production-level inference, where memory capacity and availability, not only GPU compute power, affect latency, system throughput, and overall user experience.

“Penguin Solutions operationalizes and optimizes AI inferencing by delivering the performance, scalability, and reliability required to realize fully actionable insight and discovery,” said Phil Pokorny, chief technology officer at Penguin Solutions. “Organizations must understand the factors that impact inference performance—which differ significantly from training—to productize AI and deliver accurate and fast outcomes. Whether it’s for deep research or agentic applications, we optimize infrastructure for real-world workloads and enable organizations to turn AI innovation into measurable business outcomes.”

Penguin’s MemoryAI™ KV Cache Server Matched with NVIDIA GPUs Optimizes OriginAI Solutions for Scalable AI Inference

Penguin Solutions OriginAI solutions also offer the flexibility to incorporate Penguin’s CXL-based MemoryAI KV cache server, designed to support customers’ KV strategies by expanding KV cache capacity, enabling low-latency, high-concurrency inference and extended context lengths for the most demanding applications. Use of Penguin’s MemoryAI KV cache server, which is compatible with the NVIDIA Dynamo framework, provides cost-efficiency and optimal design for the next wave of AI deployment.

OriginAI AI factory solutions also include Penguin Solutions ICE ClusterWare™ software, an intelligent management layer that transforms validated hardware into a fully-tuned AI cluster. ICE ClusterWare software delivers health monitoring and auto-remediation, to ensure sustained peak performance at scale. It also enhances data security in multi-tenant environments by isolating workloads and protecting sensitive information.

The OriginAI portfolio offers a range of configurations to address diverse customer needs. NVIDIA RTX PRO 6000-based architecture targets enterprise-class copilots, retrieval-augmented generation (RAG) systems, code assistance, and document summarization, delivering a lower acquisition cost, flexible deployment, and power-efficient performance for mid-sized models. NVIDIA B300-based architecture is designed for enterprise-wide AI platforms, long-context assistants, frontier model hosting, and agentic workloads, providing massive memory bandwidth and future-proof scalability for large, shared services.

Enterprise Inference for Financial Services, Healthcare, and Retail

OriginAI inference architectures help provide the flexibility to scale out and avoid overprovisioning by combining expert infrastructure design with meticulous in-factory builds and on-site deployment. This approach enables enterprises as well as cloud service providers (CSPs) and neoclouds to cost-efficiently deploy infrastructure tailored for use case and inference applications at scale. For example:

AI is reshaping how organizations achieve efficiency, accuracy, and innovation. Penguin Solutions has delivered solutions that address customers' inference objectives and KV strategies, helping them meet evolving demands and achieve measurable results.

To learn more, explore Penguin Solutions’ OriginAI inference solutions and/or visit booth #1031 at the NVIDIA GTC AI Conference and Expo March 16-19, 2026, in San Jose, Calif.

MemoryAI, OriginAI, and ICE ClusterWare are trademarks or registered trademarks of Penguin Solutions, Inc. or its affiliates. All other trademarks are the property of their respective owners.

About Penguin Solutions

The most transformative technological advancements are often the hardest to deploy and optimize. Penguin Solutions, the AI factory platform company, has the innovative technologies, skills, experience, and partnerships needed to turn your AI ambitions into reality.

In addition to our AI capabilities, Penguin Solutions offers memory and LED solutions serving a wide range of high-performance and specialized applications.

For more information, visit www.penguinsolutions.com.

Penguin Solutions’ OriginAI Factory Platform delivers optimized performance for AI inference with the expansion of its OriginAI portfolio with solutions that address the need for more GPU memory to solve context size and concurrency, and meet low latency demands of enterprise-scale AI inference.

Penguin Solutions’ OriginAI Factory Platform delivers optimized performance for AI inference with the expansion of its OriginAI portfolio with solutions that address the need for more GPU memory to solve context size and concurrency, and meet low latency demands of enterprise-scale AI inference.

WASHINGTON (AP) — The Supreme Court will hear arguments over the Trump administration’s push to end legal protections for people fleeing war and natural disaster from countries around the world, including Haiti and Syria.

The justices refused to immediately lift the protections for hundreds of thousands of people Monday, allowing them to live and work in the U.S. legally for now.

The court is expected to hear the case next month.

The conservative-majority court has sided with the Trump administration on the issue before and allowed the end of temporary legal status for a total of 600,000 people from Venezuela while lawsuits play out. That exposed them to potential deportation amid the administration's wider crackdown on immigration.

The Trump administration filed emergency appeals after lower courts stopped the immediate end of the program for 350,000 people from Haiti and 6,000 people from Syria.

The administration asked the court to lift those decisions and issue a broad ruling that would block courts from intervening when Homeland Security decides to end protections.

The Justice Department argued that the Department of Homeland Security has sole power over the program, which was designed to be temporary.

“Lower courts are again attempting to block major executive-branch policy initiatives in ways that inflict specific harms to the national interest and foreign relations,” Solicitor General D. John Sauer wrote in court documents.

But immigration attorneys argued that both countries are still largely in crisis and people can't return safely.

“Without a functioning government, Haiti is a nation in turmoil. Rape, kidnapping, and murder are rampant, while food, housing, and medical care are scarce,” attorneys wrote, pointing to reports that four Haitian women were recently found dead months after they were deported from the U.S.

Courts in New York and Washington, D.C., have agreed to delay the end of protections, with one finding that “hostility to nonwhite immigrants” likely played a role in the decision to end protections for Haitians. During his presidential campaign, Trump amplified false rumors that Haitian immigrants were abducting and eating dogs and cats.

Appeals courts left the decisions in place.

A total of about 1.3 million people fleeing armed conflict, natural disasters and political instability in countries around the world have been granted temporary protected status. Federal authorities have said conditions in the affected countries have improved and denied racial animus played a role.

The protections for Haitians were first granted in 2010 after a catastrophic earthquake and have been extended multiple times amid ongoing gang violence that has displaced more than a million people, according to court documents.

Protections for Syrians were first granted protected status in 2012, during a civil war that lasted for more than a decade before the fall of President Bashar Assad’s government in late 2024.

Congress created TPS in 1990 to prevent deportations to countries suffering from natural disasters, civil strife or other dangerous conditions. The designation is granted in 18-month increments by the Homeland Security secretary.

It allows people to legally live and work in the U.S., though it does not provide a path to citizenship. DHS has moved to terminate the program for people from multiple countries since Republican Donald Trump returned to the White House.

The U.S. Supreme Court as seen during a snowy day on Capitol Hill Thursday, March 12, 2026, in Washington. (AP Photo/Jose Luis Magana)

The U.S. Supreme Court as seen during a snowy day on Capitol Hill Thursday, March 12, 2026, in Washington. (AP Photo/Jose Luis Magana)

Recommended Articles