AWS Logo
Menu
 Unlock Your Amazon Bedrock Cost Optimization

Unlock Your Amazon Bedrock Cost Optimization

If you give it a lot of excess, you're going to be scanning through all that and indexing all that, which would be unnecessary costs. So that goes into kind of the data component, you know, kind of garbage in garbage out

Steph Gooch
Amazon Employee
Published May 2, 2025

Introduction:

In this episode of the "Keys to AWS Optimization" show, host Stephanie Gooch, Senior Optimization Solutions Architect Advocate, is joined by Adam Richter, a Solutions Architect on the OPTICS team, and David Tepper, CEO of Pay-i, to discuss strategies for optimizing AWS Bedrock, a managed AI service. With their combined expertise in FinOps, AI, and cost efficiency, they provide valuable insights for businesses looking to leverage generative AI while keeping costs under control.

Key Takeaways:

1. Understanding the Difference Between AIML and Generative AI: David Tepper explains the technical distinctions between traditional machine learning and the emerging field of generative AI, which involves models capable of reasoning and producing new content.
2. Navigating the AWS Bedrock Ecosystem: Adam Richter delves into the features and pricing models of AWS Bedrock, highlighting the importance of understanding the cost implications of native versus third-party AI models, as well as the usage of vector databases for knowledge bases.
3. Optimizing Bedrock with Distillation, Fine-tuning, and Latency Optimization: The guests discuss various techniques for improving the performance and cost-effectiveness of AI models, including model distillation, fine-tuning, and leveraging the latency-optimized feature in Bedrock.
4. Evaluating AI Model Performance: David Tepper shares insights on the importance of evaluating AI model performance, including considerations around availability, throughput, and the different types of latency metrics relevant to generative AI.
5. The Role of Retrieval Augmented Generation (RAG): Adam Richter and David Tepper explain the concept of RAG, which allows AI models to augment their knowledge by retrieving and incorporating external data sources, and how this feature can be leveraged in Bedrock.
Notable Quotes:
- "If you give it a lot of excess, you're going to be scanning through all that and indexing all that, which would be unnecessary costs. So that goes into kind of the data component, you know, kind of garbage in garbage out." - Adam Richter
- "We transform Gen AI spend into business value and so you can think about it as FinOps plus plus specifically for generative AI." - David Tepper

Call to Action:

If you're interested in learning more about optimizing your AI costs and leveraging the capabilities of AWS Bedrock, be sure to check out the on-demand recording of this episode on the AWS YouTube channel. Additionally, the guests will be presenting at the upcoming FinOpsX conference, so be sure to attend their session to dive deeper into these topics.

Any opinions in this post are those of the individual author and may not reflect the opinions of AWS.

Comments