Newsletter Subject

A re:Invent exclusive: AWS CEO Adam Selipsky to reveal a new generative AI stack

From

substack.com

Email Address

johnfurrier@substack.com

Sent On

Mon, Nov 27, 2023 11:20 PM

Email Preheader Text

An interview with AWS CEO Adam Selipsky, ahead of re:Invent, on the multifaceted impact of generativ

An interview with AWS CEO Adam Selipsky, ahead of re:Invent, on the multifaceted impact of generative AI, the role of AI chips, the OpenAI fiasco, and more                                                                                                                                                                                                                                                                                                                                                                                                                 Forwarded this email? [Subscribe here]() for more [A re:Invent exclusive: AWS CEO Adam Selipsky to reveal a new generative AI stack]( An interview with AWS CEO Adam Selipsky, ahead of re:Invent, on the multifaceted impact of generative AI, the role of AI chips, the OpenAI fiasco, and more [JohnFurrier.substack.com]( Nov 27   [READ IN APP](   As the rush to generative artificial intelligence only gains more momentum every day, Amazon Web Services Inc. Chief Executive Adam Selipsky think there's one core capability to ensure it fulfills its promise: adaptability. "With all of the different directions that generative AI, and AI, are going to go in, the key characteristic that people need to have is adaptability," Selipsky said in an exclusive interview with SiliconANGLE before AWS re:Invent in Las Vegas. "AI will branch out in various directions, but when it does so simultaneously, being adaptive will be key to who wins."   Selipsky's quote encapsulates the essence of what promises to be a dynamic re:Invent, the premier cloud computing event of the year. A mere year ago, at re:Invent 2022, generative AI was scarcely a topic of discussion, coming just days after OpenAI introduced its revolutionary ChatGPT chatbot that has since taken the world by storm. In just one year, gen AI has become central to the tech world's evolution. AWS is now at a critical juncture, facing perhaps its most significant challenge and transformation in cloud computing since establishing itself as an industry leader. In a comprehensive and far-reaching dialogue, Selipsky and I engaged in an open discussion about the multifaceted impact of generative AI. We touched upon various aspects including its influence on business, the role of silicon chips, the evolution of the technology stack, and the competitive landscape. The CEO also delved into how AWS is navigating the innovations brought about by generative AI, exploring the implications not just for AWS itself but also for its customers and competitors in the industry. The elephant in the room: the OpenAI situation The recent drama surrounding OpenAI and Microsoft Corp. has raised global concerns about AI's risks, safety, security, performance, cost and even viability. My interview with Selipsky concluded just an hour before CEO Sam Altman's departure from OpenAI, preventing me from directly asking about this development. However, in a follow-up, Selipsky shared his thoughts on the situation (see full quote below) and the recent drama, emphasizing the importance of customer choice in AWS' generative AI strategy, a principle established nearly a year ago. "It's essential for companies to have access to a variety of models; no single model or provider should dominate," he said. "The recent events have reinforced the soundness of AWS' chosen approach." Selipsky said he believes that "reliable models and dependable providers are essential, as are cloud providers who offer choices and are committed to technology that supports these choices." AWS, he contends, "has advocated and implemented these principles for over 17 years, and recent developments in generative AI have further highlighted their importance." When discussing the competition, Selipsky refrained from naming Microsoft directly but made it clear whom he was referring to. He expressed his astonishment at how "other providers" released early versions of AI offerings without comprehensive security models. “It boggles my mind how other providers have released early versions of their AI offerings essentially without a security model,” he said. For context I've included Selipsky's full comment on the OpenAI situation: “We have been saying since we started laying out our generative AI strategy, almost a year ago, how important it is for customers to have choice. Customers need multiple models - there is no one model and no one model provider to rule them all. The events of the last few days have shown the validity and the usefulness behind the strategy that AWS has chosen. Customers need models they can depend on. Customers need model providers that are going to be dependable business partners. And customers need cloud providers who offer choice and offer the technology that enables choice. Those cloud providers need to want to offer choice as opposed to being driven to provide choice. The importance of these attributes, which AWS has been talking about and acting on for more than 17 years, has become abundantly clear to customers in relation to generative AI.” Generative AI business growth While generative AI dominates the conversations among enterprises and developers, cloud growth has slowed as customers continue to "rightsize" their cloud spend to adjust to economic pressures and prepare to invest in new generative AI infrastructure and applications.   Selipsky notes that most cost optimization is complete, with increasing activity around new workloads such as generative AI. Industry observers are now analyzing the dynamics of cost optimization versus AWS's handling of long-term contracts and their financial impact. To win the battle for cloud supremacy, AWS must keep growing rapidly, maintain a strong cloud AI product portfolio and increase its market share. As the market transitions from gen AI experimentation to large-scale workloads, significant growth is expected to return. "Growth rates have stabilized, and we're cautiously optimistic," Selipsky explains. He continues, "many customers have completed their cost optimization, and we're hopeful for increased growth. Looking at the mid to long term, we feel very optimistic about the outlook for strong AWS growth. We're still in the early days of the cloud". Building generative AI apps Questions arise about building and deploying generative AI models and applications in a rapidly evolving AI landscape. AWS' answer is a new foundational architecture for gen AI, comprising a three-layer technology stack. This stack includes an infrastructure layer, a foundation model service layer and an AI applications layer in an attempt to make it easy for customers to innovate at all three levels. Dave Vellante, chief analyst at theCUBE Research, SiliconANGLE Media's market research arm, says customers seek to build systems or platforms leveraging chips, data, foundational models and cloud services across all three of these layers to foster more productive and adaptable gen AI applications. Selipsky points out the similarities between the AWS generative AI tech stack and the layers of other AWS cloud services. "Our unique generative AI stack offers customers advantages over alternative clouds," he says. "Not all competitors have chosen to innovate at every layer, and one must wonder how long it will take them to catch up." It’s clear that AWS’ main strategy on winning AI cloud supremacy is grounded on both continuing to enhance its cloud infrastructure and building out a unique gen AI tech stack for the market. Selipsky believes that the benefits of the new AI stack from AWS will be focused on model choice, silicon cost and performance, and AI developer success in building, training and running foundational models of a new generation of AI applications.   Silicon for gen AI AWS' track record in silicon technology, demonstrated by the Nitro hypervisor and multiple generations of Graviton, Trainium and Inferentia chips, gives it a significant edge in the evolution of cloud and generative AI. Selipsky explained the tangible benefits of these innovations, noting the importance of balancing computational power with cost considerations. "Gen AI workloads are incredibly compute-intensive, so price performance is absolutely vital," he said. This approach positions AWS as a major cloud provider in the generative AI market. AWS should be expected to continue the tradition of making big strides in chip technology with the Graviton series, with Graviton 3 in the market for over a year. Each generation of AWS' custom chips has delivered industry-leading price performance and energy efficiency. According to Selipsky, it's extremely important that AWS keep advancing general-purpose computing and innovating at the silicon layer in technology in compute and storage.  Model choice with silicon differentiation In the race for superior AI foundational models, AWS' strategic partnership with Anthropic, which includes OpenAI alumni, is a significant component of its Foundation Model service layer. The collaboration with Anthropic, backed by a substantial investment of $1.25 billion, potentially growing to $4 billion, contrasts with OpenAI's partnership with Microsoft. AWS offers multiple models through its Amazon Bedrock foundation model service, supported by generations of infrastructure and silicon capability. Selipsky highlighted the strategic importance of Anthropic's role in AWS' Bedrock service, noting the startup's insights in making chips faster, more efficient and more cost-effective for demanding AI workloads. "Anthropic's need for large amounts of compute capacity to train their models can help us enhance our Trainium and Inferentia developments," he explained. The CEO also hinted at exclusive customization features around AWS services, available only through Amazon Bedrock and Anthropic's first-party offering. "There's going to be important fine-tuning and customization features, which will only be available for limited periods of time on Amazon Bedrock and through Anthropic’s first party offering,” explained Selispky. "Not through any other channel." Discussing the insights from the AWS-Anthropic chip development, Selipsky revealed, "we'll gain insights to make chips more performant, increase throughput, and enhance efficiency and performance." This collaboration not only strengthens AWS' position in generative AI but also increases the competition with the other cloud players.  With the silicon edge that’s paired with foundation models, AWS hopes to gain insights into how models are built at scale and then offering new silicon services that only they can offer customers.  Selipsky explained the depth of the Anthropic partnership: "They will be developing their future models on AWS, they've named AWS their primary cloud provider for mission-critical workloads," Selipsky said. "And they're going to be running the majority of their workloads on AWS."  The GPU conundrum Nvidia's dominance in GPUs significantly influences the gen AI cloud computing market. However, Selipsky pointed out that AWS' philosophy embraces both custom silicon and strong partnerships with companies such as Nvidia Corp. "Amazon loves the word 'AND'," he said. "We have a great and strong relationship with Nvidia, AND AWS is a leading host for GPU-based capacity for generative AI." He addresses the complexity of building effective gen AI infrastructures, saying it involves more than just GPUs. It's about creating clusters that are highly performant, reliable, cost-effective, and energy-efficient. The challenge lies in integrating silicon chips with critical infrastructure services like networking, storage, and scalability. The discussion around chip technology in the AI sector extends far beyond the chips themselves. It encompasses a broader cloud ecosystem involving infrastructure, data services, and cloud services, all intricately connected to how chips function. The use of silicon chips is just one piece of a complex puzzle in managing varied and sophisticated AI workloads. The real challenge in generative AI lies in integrating these chips with critical infrastructure services like networking, storage, and the scalability of clusters. This integration is vital for the future of generative AI as workloads grow in diversity and complexity.  Selipsky thinks customers recognize that it's not just about having chips, but also about having highly performant services around the chips, such as networking inside the clusters. “We've seen customers go and investigate it [their own GPU clusters] and then come running back to us saying, you know, having chips is great, but it doesn't actually work,” he said.  Future of foundation models: openness and choice In the quest generative AI success, the true power lies not just in the training of models but more significantly in the phase of inference, where real value and insights are extracted from data. This distinction between training and inference is a pivotal aspect of AI's practical application. At the recent KubeCon conference, it was noted that inference is the new web app, meaning inference will be the key ingredient for leveraging data to power new web apps, which will be AI applications. "Data is the key thing," Selipsky agreed. "First, you do need really great models. People are going to want models from multiple providers." That reality speaks to the industry’s debate on open models that offer a diverse range of models catering to various needs and AI workload use cases. When asked if there will be one model to rule them all, Selipsky swiftly dispelled the notion: 'There's not going to be one model to rule the world," he said. "The idea that there'd be one company or one model is preposterous to me.' Selipsky envisions a future in gen AI where the heterogeneity of models from multiple suppliers, in different sizes and capabilities, is key. 'There will be multiple models," he said. "There will be some very large, highly capable, general-purpose models. And there will also be smaller, specialized models." He points out that these smaller models have two major advantages: They can be fine-tuned for specific queries, and they offer better price-performance ratios. This approach reflects a nuanced understanding of the different needs of AI applications. 'There may be cases where the much larger model will deliver 5% better answers at twice the price," Selipsky noted. "And for certain questions, you need better answers than those provided by the large model. For many situations, you're more than happy to pay half and get an answer that is almost as good as the large models. It’s a tradeoff." The foundational model layer for gen AI will create more opportunities for creative value creation when models are chosen for the right task and with the right infrastructure to support them. As a result, there is a growing interest and rapid experimentation in developing generative AI platforms and applications, yet the transition to actual production workloads remains in its early stages. According to Selipsky, this scenario will evolve rapidly as more companies master the art of constructing, deploying and operating gen AI infrastructure and development environments that effectively use their data. Selipsky emphasized again the importance of adaptability in generative AI: "With AI branching out in various directions, being adaptive will be key to who wins when these developments occur simultaneously," he said. "The benefits of gen AI come when things are easy to use and simple. It's this kind of experimentation that is only possible if it's simple to move back and forth between models. That's why the choice we're offering at the middle of the stack is so important.'" Deploying and applying gen AI Although the current focus is on gen AI model training, a significant shift toward inference is expected, emphasizing the ability to interpret and utilize data effectively. “You need training and you need inference," Selipsky said. "Today, there's a lot of focus on training because it's early and we're building models to deploy generative AI applications. Once this happens, the emphasis and resource allocation in gen AI will shift more toward inference.”  Enterprises are increasingly recognizing that their data is a key competitive advantage. As gen AI transforms the application layer and user interface, the approach to data, viewed as intellectual property and a competitive edge, needs to be rethought. Establishing an effective data strategy and a robust data platform is essential, especially for those excelling in gen AI. The focus is now shifting toward data management, set to become as crucial as model management and model security. This shift represents a new approach to data management, altering the traditional methods. Changes in the technology stack will have significant implications, particularly for data, which in turn affects applications. “Great models will be built, and they are going to be important," Selipsky said. "Both you and your competitors will have equal access to these models. However, your data will be what sets you apart. Our customers are rapidly advancing in generative AI because they already have a strong data strategy implemented on AWS. The more the generative AI understands or has data about how your company operates, including your code bases and libraries, the more powerful it's going to become for developers.” An effective data strategy also requires a thorough understanding of the available data, ensuring it's harmonized and usable across various applications. Key challenges include the need for multiple database engines rather than just one, and a range of analytics services, each with extensive capabilities. Governance is crucial as well, as it ensures companywide awareness of available data and manages permissions appropriately. This means that different individuals have varying access permissions to different data sets, and these permissions must be consistently upheld across all services used. The bottom line A vision for the future of generative AI and cloud computing emerged from my interview with Selipsky, highlighting AWS' pivotal role in this rapidly evolving field. Selipsky's insights reflect a world where adaptability and innovation are key, particularly in the realms of AI model training, inference and data utilization.  The industry, navigating through rapid evolution, acknowledges the crucial role of generative AI adaptability and innovation, elements that AWS is demonstrating with the development of cutting-edge technologies such as the Graviton chips and specialty chips such as Trainium and Interentia. AWS is making a big bet on offering a unique three-layer generative AI technology stack that offers a diverse array of AI models and platforms, strategic partnerships, best price-performance technology, and commitment to choice.   Can AWS continue to deliver the value to the needs of customers, partners, and developers facing diverse challenges and opportunities of the rapidly advancing tech world? Selipsky thinks it can and will.  “We're very, very bullish on the business long-term,” he said. Photo: AWS John @Furrier Newsletter is free today. But if you enjoyed this post, you can tell John @Furrier Newsletter that their writing is valuable by pledging a future subscription. You won't be charged unless they enable payments. [Pledge your support](   [Like]( [Comment]( [Restack](   © 2023 John Furrier 989 Commercial Street, Palo Alto, CA 94303 [Unsubscribe]() [Get the app]( writing]()

EDM Keywords (272)

year writing world workloads word wins win well want vital vision variety value valuable validity use twice transition transformation trainium training train tradition tradeoff topic time three thoughts things technology talking take supports support strategy storm still startup stack stabilized soundness slowed simultaneously simple similarities siliconangle significantly shown shift sets selipsky scenario scarcely scale scalability says said rush running rule room role rightsize reveal result relation reinforced referring realms range race providers provider provided promises productive principles preposterous prepare powerful post possible points pledging phase performance partnership paired outlook optimistic opposed opportunities openai one offers offering offer nvidia notion noted needs need navigating models mind middle mid means may market making make majority made lot long libraries layers last know kind key involves investigate invest invent interview interpret integration integrating insights innovation innovating innovate infrastructure influence inference industry increase important importance implications implemented idea hour hopeful highlighted heterogeneity harmonized happy happens handling grounded great gpus good going go get generations generation gains future fulfills foster forwarded forth follow focused focus feel extracted expressed explained experimentation expected excelling evolution events essential essence ensure enjoyed enhance engaged encompasses emphasis elephant efficient easy early dynamics dynamic driven dominate dominance diversity distinction discussing development developing developers depth depend departure demonstrating deliver debate days data customers crucial create continuing continue context contends compute comprehensive complexity completed complete competitors competition companies committed commitment collaboration clusters cloud clear chosen choice chips catch cases capabilities business bullish built building branch boggles benefits believes become battle aws available attributes attempt astonishment asked art approach applications app apart anthropic answer analyzing also already almost ai advocated adjust addresses adaptive adaptability acting access ability

substack.com

JohnFurrier.substack.com from John @Furrier Newsletter

Marketing emails from substack.com

View More
Sent On

08/06/2024

Sent On

07/06/2024

Sent On

07/06/2024

Sent On

07/06/2024

Sent On

07/06/2024

Sent On

07/06/2024

Email Content Statistics

Subscribe Now

Subject Line Length

Data shows that subject lines with 6 to 10 words generated 21 percent higher open rate.

Subscribe Now

Average in this category

Subscribe Now

Number of Words

The more words in the content, the more time the user will need to spend reading. Get straight to the point with catchy short phrases and interesting photos and graphics.

Subscribe Now

Average in this category

Subscribe Now

Number of Images

More images or large images might cause the email to load slower. Aim for a balance of words and images.

Subscribe Now

Average in this category

Subscribe Now

Time to Read

Longer reading time requires more attention and patience from users. Aim for short phrases and catchy keywords.

Subscribe Now

Average in this category

Subscribe Now

Predicted open rate

Subscribe Now

Spam Score

Spam score is determined by a large number of checks performed on the content of the email. For the best delivery results, it is advised to lower your spam score as much as possible.

Subscribe Now

Flesch reading score

Flesch reading score measures how complex a text is. The lower the score, the more difficult the text is to read. The Flesch readability score uses the average length of your sentences (measured by the number of words) and the average number of syllables per word in an equation to calculate the reading ease. Text with a very high Flesch reading ease score (about 100) is straightforward and easy to read, with short sentences and no words of more than two syllables. Usually, a reading ease score of 60-70 is considered acceptable/normal for web copy.

Subscribe Now

Technologies

What powers this email? Every email we receive is parsed to determine the sending ESP and any additional email technologies used.

Subscribe Now

Email Size (not include images)

Font Used

No. Font Name
Subscribe Now

Copyright © 2019–2024 SimilarMail.