What are you looking for ?
Infinidat
Articles_top

MemVerge: Memory Machine V1.2 and Programs for Intel 3rd Gen Xeon Scalable Processors and Octane Persistent Memory 200 Series 

Software delivers big memory performance and capacity leveraging up to 40 cores in 3rd Gen Xeon Scalable processors and up to 6TB of capacity per socket with Optane persistent memory 200 series.

MemVerge, Inc. announced the Memory Machine software version 1.2.

Click to enlarge

Memverge Memory Machine Dashboard Screenshot

The software delivers big memory performance and capacity leveraging up to 40 cores in 3rd Gen Intel Xeon Scalable processors (code named Ice Lake) and up to 6TB of capacity per socket with Optane persistent memory 200 series. The company also announced its membership in the CXL Consortium, and five Big Memory Labs at Arrow, Intel, MemVerge, Penguin Computing, and WWT that are now equipped and available for big  memory demonstrations, PoC testing, and software integration.

Memory Machine v1.2 is designed to allow application vendors and end-users to take full advantage of Intel’s latest Xeon Scalable processor and Optane memory technology,” said Charles Fan, CEO. “We started by providing access to new levels of performance and capacity without requiring changes to applications.

Big memory is bigger, faster, and more available with Memory Machine v1.2
Pioneered by the company’ software, big memory software makes 100% use of available memory capacity while providing new operational capabilities to memory-centric workloads such as virtualized cloud infrastructure, in-memory databases, genomics, and animation/VFX. Memory Machine v1.2 adds the following capabilities to maket big memory bigger, faster, and more available to a broader set of applications:

  • Support for 3rd Gen Xeon Scalable processors – from 8 to 40 cores

  • Support for Optane Persistent Memory 200 Series – 32% more bandwidth and up to 6TB per socket vs. previous gen

  • Centralized memory management – configuration, monitoring, management, and alerts for DRAM and PMEM across the data center

  • Redis and Hazelcast Cluster HA – coordinated in-memory snapshots among members of in-memory database clusters enables instant recovery of the entire cluster.

  • Support for SQL Server on Linux – Double the performance for OLTP with the same memory cost by leveraging 100% of DRAM and PMEM capacity and proprietary memory tiering technology.

  • Support for KVM hypervisors – Memory Machine 1.2 supports advanced memory management for QEMU-KVM, enabling utilization of 100% of DRAM+PMEM capacity, dynamic tuning of DRAM:PMEM ratio per VM, and minimizes performance degradation caused by noisy neighbors.

  • Acceleration of Genomic Analytics – Memory Machine v1.1 has proven to accelerate overall single-cell analytics pipelines by 60% by eliminating storage I/O. Read and write testing with v1.2 is expected to improve those results.

  • Animation and VFX HA and efficiency – Autosave and in-memory snapshots allow animation and VFX apps to provide ‘Time Machine’ capabilities that allow artists to share workspaces instantly and recover from crashes ins.

According to Mark Wright, technology manager, Chapeau Studios, “Initially, we opened a poly-dense scene in Maya and it took two-and-a-half minutes. Then, we opened a scene from a snapshot we’d taken with Memory Machine and it took 8s. In addition to opening exponentially faster, another benefit of the Memory Machine snapshot is that it gets an artist right to the spot in the application where they were when they created a snapshot. There’s no need to repopulate the entire application.

Apps Sizzle on Ice Lake in Testing
Independent Lab StorageReview.com pulled together a server configured with Optane 200 Series Persistent Memory, Gen 3 Xeon Scalable CPUs, and Memory Machine software from the company. They performed bulk insert and read tests with kdb+, as well as Redis Quick Recovery with ZeroIO snapshot and Redis Clone with ZeroIO snapshot.

In summary, the configuration with 200 Series PMEM, Intel Gen 3 Xeon Scalable CPUs, and Memory Machine software demonstrated 2x read performance and 3x write performance. Full review.

Click to enlarge

Memverge Makes Big Memory Apps Sizzle

According to Kevin O’Brien, lab director, StorageReview.com, “The real benefits of PMEM show up when you can leverage it at the byte level with the appropriate software. In many cases, application developers like SAP, tune their application to be able to leverage PMEM. While that works for some applications, there’s another option. Leverage a software-defined solution that’s built from the ground up to help businesses leverage all of the performance and persistence benefits PMEM 200 offers. To test this latest gen of PMEM, that’s exactly what we did.”

Big memory labs accelerate technology evaluation and integration
Big memory makes use of Optane persistent memory and the company’s Memory Machine software. For IT organizations and vendors that need quick access to a big memory environment, 5 big memory labs are available for demonstrations, PoC testing, and software integration. See Big Memory Lab page for information about capabilities and scheduling time in each lab: Arrow, Intel, MemVerge, Penguin Computing, StorageReview.com, and WWT.

Big memory opportunity
DRAM was invented in 1969. Over 50 years later, it remains expensive, scarce, and volatile. Clearly DRAM’s speed of evolution cannot keep pace with the demand of modern applications that must process large quantities of data and deliver results in real-time. Fortunately, the invention of Intel 3D XPoint technology breathed new life into the aging segment, and marked the advent of the age of big memory computing.

By 2024, almost a quarter of all data created will be real-time data, and two-thirds of Global 2000 corporations will have deployed at least one real-time application that is considered mission-critical, according to IDC.

These next-gen applications frequently employ big data analytics and AI/ML. The real-time workloads are found in many industries on-premises and in the cloud. Examples include in-memory databases and fraud analytics in financial services, customer profiling in social media, recommendation engines in retail, 3D animation in M&E, genomics in health sciences, and security forensics.

The result is explosive adoption of big memory computing designed for big and fast data. IDC estimates the market opportunity for persistent memory will grow from $65 million to $2.6 billion by 2023. Coughlin & Associates estimates that revenue for persistent memory will reach $25 billion by 2030, equal to revenue for DRAM.

Read also :
Articles_bottom
AIC
ATTO
OPEN-E