aster.cloud aster.cloud
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
aster.cloud aster.cloud
  • /
  • Platforms
    • Public Cloud
    • On-Premise
    • Hybrid Cloud
    • Data
  • Architecture
    • Design
    • Solutions
    • Enterprise
  • Engineering
    • Automation
    • Software Engineering
    • Project Management
    • DevOps
  • Programming
    • Learning
  • Tools
  • About
  • Cloud-Native

Here’s What AWS Revealed About Its Generative AI Strategy At re:Invent 2023

  • aster.cloud
  • December 8, 2023
  • 6 minute read

AWS CEO Adam Selipsky and other top executives at the company’s annual re:Invent conference revealed updates and new products aimed at challenging offerings from Microsoft, Oracle, Google, and IBM.

At AWS’ annual re:Invent conference this week, CEO Adam Selipsky and other top executives announced new services and updates to attract burgeoning enterprise interest in generative AI systems and take on rivals including Microsoft, Oracle, Google, and IBM. 


Partner with aster.cloud
for your next big idea.
Let us know here.



From our partners:

CITI.IO :: Business. Institutions. Society. Global Political Economy.
CYBERPOGO.COM :: For the Arts, Sciences, and Technology.
DADAHACKS.COM :: Parenting For The Rest Of Us.
ZEDISTA.COM :: Entertainment. Sports. Culture. Escape.
TAKUMAKU.COM :: For The Hearth And Home.
ASTER.CLOUD :: From The Cloud And Beyond.
LIWAIWAI.COM :: Intelligence, Inside and Outside.
GLOBALCLOUDPLATFORMS.COM :: For The World's Computing Needs.
FIREGULAMAN.COM :: For The Fire In The Belly Of The Coder.
ASTERCASTER.COM :: Supra Astra. Beyond The Stars.
BARTDAY.COM :: Prosperity For Everyone.

AWS, the largest cloud service provider in terms of market share, is looking to capitalize on growing interest in generative AI. Enterprises are expected to invest $16 billion globally on generative AI and related technologies in 2023, according to a report from market research firm IDC.  

This spending, which includes generative AI software as well as related infrastructure hardware and IT and business services, is expected to reach $143 billion in 2027, with a compound annual growth rate (CAGR) of 73.3%.  

This exponential growth, according to IDC, is almost 13 times greater than the CAGR for worldwide IT spending over the same period.  

Like most of its rivals, particularly Oracle, Selipsky revealed that AWS’ generative strategy is divided into three tiers — the first, or infrastructure, layer for training or developing large language models (LLMs); a middle layer, which consists of foundation large language models required to build applications; and a third layer, which includes  applications that use the other two layers.

AWS beefs up infrastructure for generative AI

The cloud services provider, which has been adding infrastructure capabilities and chips since the last year to support high-performance computing with enhanced energy efficiency, announced the latest iterations of its Graviton and the Trainium chips this week.

The Graviton4 processor, according to AWS, provides up to 30% better compute performance, 50% more cores, and 75% more memory bandwidth than the current generation Graviton3 processors.

Trainium2, on the other hand, is designed to deliver up to four times faster training than first-generation Trainium chips.

These chips will be able to be deployed in EC2 UltraClusters of up to 100,000 chips, making it possible to train foundation models (FMs) and LLMs in a fraction of the time than it has taken up to now, while improving energy efficiency up to two times more than the previous generation, the company said. 

Read More  Sora

Rivals Microsoft, Oracle, Google, and IBM all have been making their own chips for high-performance computing, including generative AI workloads.

While Microsoft recently released its Maia AI Accelerator and Azure Cobalt CPUs for model training workloads, Oracle has partnered with Ampere to produce its own chips, such as the Oracle Ampere A1. Earlier, Oracle used Graviton chips for its AI infrastructure. Google’s cloud computing arm, Google Cloud, makes its own AI chips in the form of Tensor Processing Units (TPUs), and their latest chip is the TPUv5e, which can be combined using Multislice technology. IBM, via its research division, too, has been working on a chip, dubbed Northpole, that can efficiently support generative workloads.  

At re:Invent, AWS also extended its partnership with Nvidia, including support for the DGX Cloud, a new GPU project named Ceiba, and new instances for supporting generative AI workloads.

AWS said that it will host Nvidia’s DGX Cloud cluster of GPUs, which can accelerate training of generative AI and LLMs that can reach beyond 1 trillion parameters. OpenAI, too, has used the DGX Cloud to train the LLM that underpins ChatGPT.

Earlier in February, Nvidia had said that it will make the DGX Cloud available through Oracle Cloud, Microsoft Azure, Google Cloud Platform, and other cloud providers. In March, Oracle announced support for the DGX Cloud, followed closely by Microsoft.

Officials at re:Invent also announced that new Amazon EC2 G6e instances featuring Nvidia L40S GPUs and G6 instances powered by L4 GPUs are in the works.

L4 GPUs are scaled back from the Hopper H100 but offer much more power efficiency. These new instances are aimed at startups, enterprises, and researchers looking to experiment with AI.

Nvidia also shared plans to integrate its NeMo Retriever microservice into AWS to help users with the development of generative AI tools like chatbots. NeMo Retriever is a generative AI microservice that enables enterprises to connect custom LLMs to enterprise data, so the company can generate proper AI responses based on their own data.

Further, AWS said that it will be the first cloud provider to bring Nvidia’s GH200 Grace Hopper Superchips to the cloud.

Read More  Speech AI Year In Review

The Nvidia GH200 NVL32 multinode platform connects 32 Grace Hopper superchips through Nvidia’s NVLink and NVSwitch interconnects. The platform will be available on Amazon Elastic Compute Cloud (EC2) instances connected via Amazon’s network virtualization (AWS Nitro System), and hyperscale clustering (Amazon EC2 UltraClusters).

New foundation models to provide more options for application building

In order to provide choice of more foundation models and ease application building, AWS unveiled updates to existing foundation models inside its generative AI application-building service, Amazon Bedrock.

The updated models added to Bedrock include Anthropic’s Claude 2.1 and Meta Llama 2 70B, both of which have been made generally available. Amazon also has added its proprietary Titan Text Lite and Titan Text Express foundation models to Bedrock.

In addition, the cloud services provider has added a model in preview, Amazon Titan Image Generator, to the AI app-building service.

Foundation models that are currently available in Bedrock include large language models (LLMs) from the stables of AI21 Labs, Cohere Command, Meta, Anthropic, and Stability AI.

Rivals Microsoft, Oracle, Google, and IBM also offer various foundation models including proprietary and open-source models. While Microsoft offers Meta’s Llama 2 along with OpenAI’s GPT models, Google offers proprietary models such as PaLM 2, Codey, Imagen, and Chirp. Oracle, on the other hand, offers models from Cohere.

AWS also released a new feature within Bedrock, dubbed Model Evaluation, that allows enterprises to evaluate, compare, and select the best foundational model for their use case and business needs.

Although not entirely similar, Model Evaluation can be compared to Google Vertex AI’s Model Garden, which is a repository of foundation models from Google and its partners. Microsoft Azure’s OpenAI service, too, offers a capability to select large language models. LLMs can also be found inside the Azure Marketplace.

Amazon Bedrock, SageMaker get new features to ease application building

Both Amazon Bedrock and SageMaker have been updated by AWS to not only help train models but also speed up application development.

These updates includes features such as  Retrieval Augmented Generation (RAG), capabilities to fine-tune LLMs, and the ability to pre-train Titan Text Lite and Titan Text Express models from within Bedrock. AWS also introduced  SageMaker HyperPod and SageMaker Inference, which help in scaling LLMs and reducing cost of AI deployment respectively.

Read More  An Introduction To OpenTelemetry And Observability

Google’s Vertex AI, IBM’s Watsonx.ai, Microsoft’s Azure OpenAI, and certain features of the Oracle generative AI service also provide similar features to Amazon Bedrock, especially allowing enterprises to fine-tune models and the RAG capability.

Further, Google’s Generative AI Studio, which is a low-code suite for tuning, deploying and monitoring foundation models, can be compared with AWS’ SageMaker Canvas, another low-code platform for business analysts, which has been updated this week to help generation of models.

Each of the cloud service providers, including AWS, also have software libraries and services such as Guardrails for Amazon Bedrock, to allow enterprises to be compliant with best practices around data and model training.

Amazon Q, AWS’ answer to Microsoft’s GPT-driven Copilot

On Tuesday, Selipsky premiered the star of the cloud giant’s re:Invent 2023 conference: Amazon Q, the company’s answer to Microsoft’s GPT-driven Copilot generative AI assistant.   

Selipsky’s announcement of Q was reminiscent of Microsoft CEO Satya Nadella’s keynote at Ignite and Build, where he announced several integrations and flavors of Copilot across a wide range of proprietary products, including Office 365 and Dynamics 365. 

Amazon Q can be used by enterprises across a variety of functions including developing applications, transforming code, generating business intelligence, acting as a generative AI assistant for business applications, and helping customer service agents via the Amazon Connect offering. 

Rivals are not too far behind. In August, Google, too, added its generative AI-based assistant, Duet AI, to most of its cloud services including data analytics, databases, and infrastructure and application management.

Similarly, Oracle’s managed generative AI service also allows enterprises to integrate LLM-based generative AI interfaces in their applications via an API, the company said, adding that it would bring its own generative AI assistant to its cloud services and NetSuite.

Other generative AI-related updates at re:Invent include updated support for vector databases for Amazon Bedrock. These databases include Amazon Aurora and MongoDB. Other supported databases include Pinecone, Redis Enterprise Cloud, and Vector Engine for Amazon OpenSearch Serverless.

By: Anirban Ghoshal
Originally published at: InfoWorld

Source: cyberpogo.com


For enquiries, product placements, sponsorships, and collaborations, connect with us at [email protected]. We'd love to hear from you!

Our humans need coffee too! Your support is highly appreciated, thank you!

aster.cloud

Related Topics
  • AI
  • Amazon Web Services
  • Artificial Intelligence
  • AWS
You May Also Like
View Post
  • Cloud-Native
  • Multi-Cloud

Oracle Expands Multicloud Capabilities with AWS, Google Cloud, and Microsoft Azure

  • September 11, 2024
Cloud computing concept image double exposure Digitally Enhanced Smart City Concept with Cloud Computing
View Post
  • Cloud-Native
  • Computing
  • Hybrid Cloud
  • Multi-Cloud
  • Public Cloud

Make Your Business Resilient By Integrating These Best Practices Into Your Cloud Architecture

  • July 29, 2024
Huawei Cloud Cairo Region Goes Live
View Post
  • Cloud-Native
  • Computing
  • Platforms

Huawei Cloud Goes Live in Egypt

  • May 24, 2024
View Post
  • Cloud-Native
  • Computing
  • Engineering

10 Cloud Development Gotchas To Watch Out For

  • March 29, 2024
Storage Ceph
View Post
  • Cloud-Native
  • Data

The Growth Of IBM Storage Ceph – The Ideal Foundation For A Modern Data Lakehouse

  • January 30, 2024
Clouds
View Post
  • Cloud-Native
  • Platforms
  • Software Engineering

Microsoft Releases Azure Migrate Assessment Tool For .NET Application

  • January 14, 2024
View Post
  • Cloud-Native
  • Engineering
  • Platforms

Top Highlights From AWS Worldwide Public Sector Partners At Re:Invent 2023

  • December 27, 2023
View Post
  • Cloud-Native
  • Computing

Supercharging IBM’s Cloud-Native AI Supercomputer

  • December 24, 2023

Stay Connected!
LATEST
  • college-of-cardinals-2025 1
    The Definitive Who’s Who of the 2025 Papal Conclave
    • May 7, 2025
  • conclave-poster-black-smoke 2
    The World Is Revalidating Itself
    • May 6, 2025
  • 3
    Conclave: How A New Pope Is Chosen
    • April 25, 2025
  • Getting things done makes her feel amazing 4
    Nurturing Minds in the Digital Revolution
    • April 25, 2025
  • 5
    AI is automating our jobs – but values need to change if we are to be liberated by it
    • April 17, 2025
  • 6
    Canonical Releases Ubuntu 25.04 Plucky Puffin
    • April 17, 2025
  • 7
    United States Army Enterprise Cloud Management Agency Expands its Oracle Defense Cloud Services
    • April 15, 2025
  • 8
    Tokyo Electron and IBM Renew Collaboration for Advanced Semiconductor Technology
    • April 2, 2025
  • 9
    IBM Accelerates Momentum in the as a Service Space with Growing Portfolio of Tools Simplifying Infrastructure Management
    • March 27, 2025
  • 10
    Tariffs, Trump, and Other Things That Start With T – They’re Not The Problem, It’s How We Use Them
    • March 25, 2025
about
Hello World!

We are aster.cloud. We’re created by programmers for programmers.

Our site aims to provide guides, programming tips, reviews, and interesting materials for tech people and those who want to learn in general.

We would like to hear from you.

If you have any feedback, enquiries, or sponsorship request, kindly reach out to us at:

[email protected]
Most Popular
  • 1
    IBM contributes key open-source projects to Linux Foundation to advance AI community participation
    • March 22, 2025
  • 2
    Co-op mode: New partners driving the future of gaming with AI
    • March 22, 2025
  • 3
    Mitsubishi Motors Canada Launches AI-Powered “Intelligent Companion” to Transform the 2025 Outlander Buying Experience
    • March 10, 2025
  • PiPiPi 4
    The Unexpected Pi-Fect Deals This March 14
    • March 13, 2025
  • Nintendo Switch Deals on Amazon 5
    10 Physical Nintendo Switch Game Deals on MAR10 Day!
    • March 9, 2025
  • /
  • Technology
  • Tools
  • About
  • Contact Us

Input your search keywords and press Enter.