According to TapTechNews on June 2, during the ongoing keynote speech by Jensen Huang at Computex Taipei 2024, Jensen Huang introduced that generative artificial intelligence will drive the reshaping of the entire software stack and demonstrated its NIM (Nvidia Inference Microservices) cloud-native microservices.
NVIDIA believes that the 'AI factory' will set off a new industrial revolution: Taking the software industry pioneered by Microsoft as an example, Jensen Huang believes that generative AI will drive its entire stack reshaping.
To facilitate the deployment of AI services for enterprises of all sizes, NVIDIA launched NIM (Nvidia Inference Microservices) cloud-native microservices in March this year.
NIM is a set of optimized cloud-native microservices aiming to shorten the time to market and simplify the deployment of generative AI models anywhere in the cloud, data centers, and GPU-accelerated workstations. It uses industry-standard APIs to abstract the complexity of AI model development and production packaging, thereby expanding the pool of developers.
TapTechNews noticed that the official has made an NIM container for the llama3 large model, which is now online on the NVIDIA official website and is open for all users to download and deploy freely.
Computex Taipei 2024 Feature