ETFOptimize | High-performance ETF-based Investment Strategies

Quantitative strategies, Wall Street-caliber research, and insightful market analysis since 1998.


ETFOptimize | HOME
Close Window

Hewlett Packard Enterprise Debuts End-to-End AI-Native Portfolio for Generative AI

Enterprise-class co-designed software and hardware solutions from HPE and NVIDIA accelerate development and deployment of GenAI applications

Today at NVIDIA GTC, Hewlett Packard Enterprise (NYSE: HPE) announced updates to one of the industry’s most comprehensive AI-native portfolios to advance the operationalization of generative AI (GenAI), deep learning, and machine learning (ML) applications. The updates include:

  • Availability of two HPE and NVIDIA co-engineered full-stack GenAI solutions.
  • A preview of HPE Machine Learning Inference Software.
  • An enterprise retrieval-augmented generation (RAG) reference architecture.
  • Support to develop future products based on the new NVIDIA Blackwell platform.

“To deliver on the promise of GenAI and effectively address the full AI lifecycle, solutions must be hybrid by design,” said Antonio Neri, president and CEO at HPE. “From training and tuning models on-premises, in a colocation facility or the public cloud, to inferencing at the edge, AI is a hybrid cloud workload. HPE and NVIDIA have a long history of collaborative innovation, and we will continue to deliver co-designed AI software and hardware solutions that help our customers accelerate the development and deployment of GenAI from concept into production.”

“Generative AI can turn data from connected devices, data centers and clouds into insights that can drive breakthroughs across industries," said Jensen Huang, founder and CEO at NVIDIA. "Our growing collaboration with HPE will enable enterprises to deliver unprecedented productivity by leveraging their data to develop and deploy new AI applications to transform their businesses.”

Supercomputing-powered GenAI training and tuning

Announced at SC23, HPE’s supercomputing solution for generative AI is now available to order for organizations seeking a preconfigured and pretested full-stack solution for the development and training of large AI models. Purpose-built to help customers accelerate GenAI and deep learning projects, the turnkey solution is powered by NVIDIA and can support up to 168 NVIDIA GH200 Grace Hopper Superchips.

The solution enables large enterprises, research institutions, and government entities to streamline the model development process with an AI/ML software stack that helps customers accelerate GenAI and deep learning projects, including LLMs, recommender systems, and vector databases. Delivered with services for installation and set-up, this turnkey solution is designed for use in AI research centers and large enterprises to realize improved time-to-value and speed up training by 2-3X. For more information or to order it today, visit HPE's supercomputing solution for generative AI.

Enterprise-class GenAI tuning and inference

Previewed at Discover Barcelona 2023, HPE’s enterprise computing solution for generative AI is now available to customers directly or through HPE GreenLake with a flexible and scalable pay-per-use model. Co-engineered with NVIDIA, the pre-configured fine-tuning and inference solution is designed to reduce ramp-up time and costs by offering the right compute, storage, software, networking, and consulting services that organizations need to produce GenAI applications. The AI-native full-stack solution gives businesses the speed, scale and control necessary to tailor foundational models using private data and deploy GenAI applications within a hybrid cloud model.

Featuring a high-performance AI compute cluster and software from HPE and NVIDIA, the solution is ideal for lightweight fine-tuning of models, RAG, and scale-out inference. The fine-tuning time for a 70 billion parameter Llama 2 model running this solution decreases linearly with node count, taking six minutes on a 16-node system1. The speed and performance enable customers to realize faster time-to-value by improving business productivity with AI applications like virtual assistants, intelligent chatbots, and enterprise search.

Powered by HPE ProLiant DL380a Gen11 servers, the solution is pre-configured with NVIDIA GPUs, the NVIDIA Spectrum-X Ethernet networking platform, and NVIDIA BlueField-3 DPUs. The solution is enhanced by HPE’s machine learning platform and analytics software, NVIDIA AI Enterprise 5.0 software with new NVIDIA NIM microservice for optimized inference of generative AI models, as well as NVIDIA NeMo Retriever and other data science and AI libraries.

To address the AI skills gap, HPE Services experts will help enterprises design, deploy, and manage the solution, which includes applying appropriate model tuning techniques. For more information or to order it today, visit HPE’s enterprise computing solution for generative AI.

From prototype to productivity

HPE and NVIDIA are collaborating on software solutions that will help enterprises take the next step by turning AI and ML proofs-of-concept into production applications. Available to HPE customers as a technology preview, HPE Machine Learning Inference Software will allow enterprises to rapidly and securely deploy ML models at scale. The new offering will integrate with NVIDIA NIM to deliver NVIDIA-optimized foundation models using pre-built containers.

To assist enterprises that need to rapidly build and deploy GenAI applications that feature private data, HPE developed a reference architecture for enterprise RAG, available today, that is based on the NVIDIA NeMo Retriever microservice architecture. The offering consists of a comprehensive data foundation from HPE Ezmeral Data Fabric Software and HPE GreenLake for File Storage. The new reference architecture will offer businesses a blueprint to create customized chatbots, generators, or copilots.

To aid in data preparation, AI training, and inferencing, the solution merges the full spectrum of open-source tools and solutions from HPE Ezmeral Unified Analytics Software and HPE’s AI software, which includes HPE Machine Learning Data Management Software, HPE Machine Learning Development Environment Software, and the new HPE Machine Learning Inference Software. HPE’s AI software is available on both HPE’s supercomputing and enterprise computing solutions for generative AI to provide a consistent environment for customers to manage their GenAI workloads.

Next-gen solutions built on NVIDIA Blackwell platform

HPE will develop future products based on the newly announced NVIDIA Blackwell platform, which incorporates a second-generation Transformer Engine to accelerate GenAI workloads. Additional details and availability for forthcoming HPE products featuring the NVIDIA GB200 Grace Blackwell Superchip, the HGX B200, and the HGXB100 will be announced in the future.

About Hewlett Packard Enterprise

Hewlett Packard Enterprise (NYSE: HPE) is the global edge-to-cloud company that helps organizations accelerate outcomes by unlocking value from all of their data, everywhere. Built on decades of reimagining the future and innovating to advance the way people live and work, HPE delivers unique, open, and intelligent technology solutions as a service. With offerings spanning Cloud Services, Compute, High Performance Computing & AI, Intelligent Edge, Software, and Storage, HPE provides a consistent experience across all clouds and edges, helping customers develop new business models, engage in new ways, and increase operational performance. For more information, visit: www.hpe.com

1 Based on initial internal benchmarks of llama-recipes finetuning.py that tracked the average epoch time to fine-tune eight nodes at 594 seconds and 16 nodes at 369 seconds with flash attention and parameter efficient fine-tuning.

Contacts

Stock Quote API & Stock News API supplied by www.cloudquote.io
Quotes delayed at least 20 minutes.
By accessing this page, you agree to the following
Privacy Policy and Terms Of Service.


 

IntelligentValue Home
Close Window

DISCLAIMER

All content herein is issued solely for informational purposes and is not to be construed as an offer to sell or the solicitation of an offer to buy, nor should it be interpreted as a recommendation to buy, hold or sell (short or otherwise) any security.  All opinions, analyses, and information included herein are based on sources believed to be reliable, but no representation or warranty of any kind, expressed or implied, is made including but not limited to any representation or warranty concerning accuracy, completeness, correctness, timeliness or appropriateness. We undertake no obligation to update such opinions, analysis or information. You should independently verify all information contained on this website. Some information is based on analysis of past performance or hypothetical performance results, which have inherent limitations. We make no representation that any particular equity or strategy will or is likely to achieve profits or losses similar to those shown. Shareholders, employees, writers, contractors, and affiliates associated with ETFOptimize.com may have ownership positions in the securities that are mentioned. If you are not sure if ETFs, algorithmic investing, or a particular investment is right for you, you are urged to consult with a Registered Investment Advisor (RIA). Neither this website nor anyone associated with producing its content are Registered Investment Advisors, and no attempt is made herein to substitute for personalized, professional investment advice. Neither ETFOptimize.com, Global Alpha Investments, Inc., nor its employees, service providers, associates, or affiliates are responsible for any investment losses you may incur as a result of using the information provided herein. Remember that past investment returns may not be indicative of future returns.

Copyright © 1998-2017 ETFOptimize.com, a publication of Optimized Investments, Inc. All rights reserved.