Plus, Google unveils Axion, its custom Arm-based chip for data centers, at Google Cloud Next 2024. Credit: Gorodenkoff / Shutterstock Alphabet’s Google Cloud unveiled a slew of new products and services at the Google Cloud Next 2024 conference in Las Vegas, among them a program to help startups and small businesses build generative AI applications and services. The initiative brings together the Nvidia Inception program for startups and the Google for Startups Cloud Program – and expands the benefits to include cloud credits, go-to-market support, and technical expertise to help startups with their AI initiatives. [ Related: More Nvidia news and insights ] Inception is a global program that supports more than 18,000 startups, and Nvidia is looking to entice more by offering an accelerated path to using Google Cloud infrastructure with access to Google Cloud credits, offering up to $350,000 for firms focused on AI. In return, Google for Startups Cloud Program members can join Nvidia Inception and gain access to technical expertise, Nvidia Deep Learning Institute course credits, Nvidia hardware and software, and more. Inception also offers a platform called Capital Connect, which gives startups exposure to venture capital firms interested in the space. Google launches Axion custom Arm processor Google also announced a new line of Arm processors for its cloud services offerings, called Axion, making it the latest tech giant to do its own custom Arm-based chips. Amazon Web Services has offered Graviton processors since 2018, and Microsoft launched its own Arm chip, called Cobalt 100, last fall. This isn’t Google’s first foray into custom silicon. It has had the tensor processing units (TPU) for acceleration of its own workloads going back to 2015, and in 2018, it launched a video coding unit (VCU) for video transcoding. But this will be its first customer-facing custom silicon. Axion is based on Arm’s Neoverse V2 design, a data-center-oriented chip built on the ARMv9 architecture. Arm doesn’t make chips; it makes designs, and then licensees take those design and do their own customizations by adding to the basic configuration they get from Arm. Some make smart phones (Apple, Qualcomm), and others make server chips (Ampere). Google declined to comment on speeds, fees, and cores, but it did claim that Axion processors would deliver instances with up to 30% better performance than the fastest general-purpose Arm-based instances available in the cloud today, up to 50% better performance, and up to 60% better energy-efficiency than comparable current-generation x86-based instances. Axion is built on Titanium, a system of Google’s own purpose-built custom silicon microcontrollers and tiered scale-out offloads. It offloads operations like networking and security, so Axion processors can focus on computation of the workload, much like the SuperNIC offloads networking traffic from the CPU. Virtual machines based on Axion processors will be available in preview in the coming months, according to Google. AI software services updated In February, Google introduced Gemma, a suite of open models using the same research and technology used to create Google’s Gemini generative AI service. Now, teams from Google and Nvidia have worked together to accelerate the performance of Gemma with Nvidia’s TensorRT-LLM, an open-source library for optimizing LLM inference. Google Cloud also has made it easier to deploy Nvidia’s NeMo framework for building custom generative AI applications across its platform via its GKE Kubernetes engine and Google Cloud HPC Toolkit. This enables developers to jumpstart the development of generative AI models, allowing for the rapid deployment of turnkey AI products. Related content news Google Cloud Next 2024: AI networking gets a boost Google's Cross-Cloud Network service gains load-balancing, security, and network management upgrades that are geared for AI workloads in multicloud environments. By Michael Cooney Apr 10, 2024 7 mins Google Cloud Next Cloud Computing Networking news Google Cloud adds more infrastructure support for AI workloads Infrastructure enhancements targeting AI workloads include updates to compute hardware, new Nvidia GPU offerings, and storage optimization. By Anirban Ghoshal Apr 10, 2024 4 mins Generative AI Google Cloud Next Google Cloud Platform news Google enhances Distributed Cloud with AI search, flexible storage Other updates to the service includes new racks, Apigee integration, and survivability features. By Anirban Ghoshal Apr 09, 2024 3 mins Generative AI Google Cloud Next Google Cloud Platform PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe