Remove Architecture Remove AWS Remove Examples
article thumbnail

Building Resilient Public Networking on AWS: Part 4

Xebia

Region Evacuation with static anycast IP approach Welcome back to our comprehensive "Building Resilient Public Networking on AWS" blog series, where we delve into advanced networking strategies for regional evacuation, failover, and robust disaster recovery. Find the detailed guide here.

AWS 130
article thumbnail

Multi-LLM routing strategies for generative AI applications on AWS

AWS Machine Learning - AI

For example, a marketing content creation application might need to perform task types such as text generation, text summarization, sentiment analysis, and information extraction as part of producing high-quality, personalized content. An example is a virtual assistant for enterprise business operations.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Build and deploy a UI for your generative AI applications with AWS and Python

AWS Machine Learning - AI

AWS provides a powerful set of tools and services that simplify the process of building and deploying generative AI applications, even for those with limited experience in frontend and backend development. The AWS deployment architecture makes sure the Python application is hosted and accessible from the internet to authenticated users.

article thumbnail

Using Amazon Q Business with AWS HealthScribe to gain insights from patient consultations

AWS Machine Learning - AI

During re:Invent 2023, we launched AWS HealthScribe , a HIPAA eligible service that empowers healthcare software vendors to build their clinical applications to use speech recognition and generative AI to automatically create preliminary clinician documentation.

AWS 113
article thumbnail

Introducing AWS MCP Servers for code assistants (Part 1)

AWS Machine Learning - AI

Were excited to announce the open source release of AWS MCP Servers for code assistants a suite of specialized Model Context Protocol (MCP) servers that bring Amazon Web Services (AWS) best practices directly to your development workflow. This post is the first in a series covering AWS MCP Servers.

AWS 119
article thumbnail

Build a multi-tenant generative AI environment for your enterprise on AWS

AWS Machine Learning - AI

It also uses a number of other AWS services such as Amazon API Gateway , AWS Lambda , and Amazon SageMaker. It contains services used to onboard, manage, and operate the environment, for example, to onboard and off-board tenants, users, and models, assign quotas to different tenants, and authentication and authorization microservices.

article thumbnail

PEFT fine tuning of Llama 3 on SageMaker HyperPod with AWS Trainium

AWS Machine Learning - AI

To simplify infrastructure setup and accelerate distributed training, AWS introduced Amazon SageMaker HyperPod in late 2023. In this blog post, we showcase how you can perform efficient supervised fine tuning for a Meta Llama 3 model using PEFT on AWS Trainium with SageMaker HyperPod. You can also customize your distributed training.

AWS 106