AWS Compute Blog

Category: Amazon Bedrock

Architecture diagram showing AWS Lambda invoking Amazon Bedrock using the InvokeModel API call.

Designing Serverless Integration Patterns for Large Language Models (LLMs)

This post is written by Josh Hart, Principal Solutions Architect and Thomas Moore, Senior Solutions Architect This post explores best practice integration patterns for using large language models (LLMs) in serverless applications. These approaches optimize performance, resource utilization, and resilience when incorporating generative AI capabilities into your serverless architecture. Overview of serverless, LLMs and example […]

Calendar

Serverless ICYMI Q2 2024

Welcome to the 26th edition of the AWS Serverless ICYMI (in case you missed it) quarterly recap. Every quarter, we share all the most recent product launches, feature enhancements, blog posts, webinars, live streams, and other interesting things that you might have missed! In case you missed our last ICYMI, check out what happened last […]

2024 Q1 calendar

Serverless ICYMI Q1 2024

Welcome to the 25th edition of the AWS Serverless ICYMI (in case you missed it) quarterly recap. Every quarter, we share all the most recent product launches, feature enhancements, blog posts, webinars, live streams, and other interesting things that you might have missed! In case you missed our last ICYMI, check out what happened last […]

Building a serverless document chat with AWS Lambda and Amazon Bedrock

This post is written by Pascal Vogel, Solutions Architect, and Martin Sakowski, Senior Solutions Architect. Large language models (LLMs) are proving to be highly effective at solving general-purpose tasks such as text generation, analysis and summarization, translation, and much more. Because they are trained on large datasets, they can use a broad generalist knowledge base. […]