What are you looking for ?
Advertise with us
RAIDON

Dell Technologies World: Pliops Unveils AI Performance Enhancements

XDP LightningAI delivers more than 2.5X end-to-end performance boost for NVIDIA Dynamo and vLLM production stack, maximizing GPU server efficiency

At Dell Technologies World 2025 in Las Vegas, NV,  Pliops Ltd. is set to showcase its XDP LightningAI solution.

Pliops Dtw 2025 Pr Lightningai Fusionx 1

In collaboration with Dell OEM Solutions, Pliops is redefining AI performance possibilities in Booth #568. The company’s XDP LightningAI solutions enable businesses to achieve up to 4X more transactions/server – tackling rack-level power constraints while also simplifying deployment and lowering costs.

At its booth, the firm will demonstrate how its XDP LightningAI solution – running on NVIDIA Dynamo and the vLLM Production Stack – delivers more than 2.5X end-to-end (E2E) performance improvements. This breakthrough redefines efficiency and significantly lowers the TCO for enterprise AI deployments.

Unlike traditional approaches that require DRAM, complex network storage, or additional indexing layers, XDP LightningAI simplifies AI infrastructure with a streamlined, single-namespace architecture – eliminating metadata overhead and ensuring seamless integration with vLLM.

The demand for scalable, high-performance AI infrastructure has never been greater,” said Ido Bukspan, CEO, Pliops. “Through our collaboration with Dell OEM Solutions, XDP LightningAI is redefining AI infrastructure efficiency. With XDP LightningAI, there is no DRAM. No network storage complexity. No indexing. No metadata. Just one unified data flow – for maximum efficiency with minimal deployment friction.”

Pliops Dtw 2025 Pr Lightningai Fusionx 2

HBM-Class Performance, Without HBM Costs
XDP LightningAI delivers performance on par with HBM – without the cost or supply constraints. Leveraging Pliops’ storage architecture, XDP LightningAI provides GPUs with an abundant supply of HBM-level memory, unlocking unprecedented efficiency for AI inference workloads. Enterprises gain HBM-class performance without the need for expensive memory upgrades – maximizing existing infrastructure investments.

Rack-Level Power Efficiency
As AI workloads grow in complexity, power constraints at the rack level have become a major bottleneck for enterprises scaling their GPU infrastructure. XDP LightningAI directly addresses this challenge by reducing power consumption per inference task, enabling organizations to:

  • Achieve up to 4X more transactions/GPU server while staying within existing power budgets.
  • Optimize power efficiency across AI inference workloads, reducing cooling and infrastructure costs.
  • Extend the lifespan of AI deployments by maximizing performance-per-watt, ensuring sustainable scalability.

Cost Savings and Dollar/Token Efficiency
XDP LightningAI delivers significant cost savings compared to traditional AI infrastructure solutions by accelerating transactions/GPU server. XDP LightningAI directly improves dollar/token efficiency, allowing enterprises to process more tokens per dollar spent. This translates into lower inference costs, making high-scale AI workloads financially sustainable without compromising performance.

Pliops FusionX stack

Click to enlarge

Pliops Fuxionx Stack Scheme

At the core of XDP LightningAI is the Pliops FusionX stack, a breakthrough technology that reduces cost, power consumption and computational overhead by optimizing LLM inference workflows.

Traditional LLM inference systems repeatedly process the same context for each request, leading to wasted compute cycles, excessive power consumption and unnecessary GPU overhead. FusionX eliminates this inefficiency by enabling context reuse, ensuring that each context is processed only once.

By reducing computational overhead, FusionX allows enterprises to scale AI inference workloads effortlessly and efficiently, accelerating performance while lowering infrastructure costs.

Exclusive Demos and Industry Validation
Adding to the excitement surrounding the show, StorageReview will publish the results from their lab testing, further showcasing the real-world performance advantages of XDP LightningAI.

Commented Brian Beeler, CEO, StorageReview, “As large language models continue to scale, so do the infrastructure challenges around inference performance and efficiency. Pliops XDP LightningAI addresses a critical pain point by enabling fast, scalable KV cache offload without compromise. Our benchmarking with Pliops and NVIDIA Dynamo demonstrates how this solution can dramatically improve GPU throughput, bringing a new level of efficiency to real-world AI deployments of any scale.”

Read also :
Articles_bottom
ExaGrid
AIC
Teledyne
ATTO
OPEN-E