Follow BigDATAwire:

December 4, 2024

AWS Advances Amazon Bedrock with AI Safeguards and Workflow Tools

LAS VEGAS, Dec. 4, 2024 — At AWS re:Invent, Amazon Web Services, Inc. (AWS) announced new capabilities for Amazon Bedrock, a fully managed service for building and scaling generative artificial intelligence (AI) applications with high-performing foundation models.

Credit: Michael Vi/Shutterstock

These newly announced capabilities help customers prevent factual errors due to hallucinations, orchestrate multiple AI-powered agents for complex tasks, and create smaller, task-specific models that can perform similarly to a large model at a fraction of the cost and latency.

  • Automated Reasoning checks is the first and only generative AI safeguard that helps prevent factual errors due to model hallucinations, opening up new generative AI use cases that demand the highest levels of precision.
  • Customers can use multi-agent collaboration to easily build and orchestrate multiple AI agents to solve problems together, expanding the ways customers can apply generative AI to address their most complex use cases.
  • Model Distillation empowers customers to transfer specific knowledge from a large, highly capable model to a smaller, more efficient one that can be up to 500% faster and 75% less expensive to run.
  • Tens of thousands of customers use Amazon Bedrock today, with Moody’s, PwC, and Robin AI among those leveraging these new capabilities to cost-effectively scale inference and push the limits of generative AI innovation.

“With a broad selection of models, leading capabilities that make it easier for developers to incorporate generative AI into their applications, and a commitment to security and privacy, Amazon Bedrock has become essential for customers who want to make generative AI a core part of their applications and businesses,” said Dr. Swami Sivasubramanian, vice president of AI and Data at AWS. “That is why we have seen Amazon Bedrock grow its customer base by 4.7x in the last year alone. Over time, as generative AI transforms more companies and customer experiences, inference will become a core part of every application. With the launch of these new capabilities, we are innovating on behalf of customers to solve some of the top challenges, like hallucinations and cost, that the entire industry is facing when moving generative AI applications to production.”

Automated Reasoning Checks Prevent Factual Errors Due to Hallucinations

While models continue to advance, even the most capable ones can hallucinate, providing incorrect or misleading responses. Hallucinations remain a fundamental challenge across the industry, limiting the trust companies can place in generative AI. This is especially true for regulated industries, like healthcare, financial services, and government agencies, where accuracy is critical, and organizations need to audit to make sure models are responding appropriately. Automated Reasoning checks is the first and only generative AI safeguard that helps prevent factual errors due to hallucinations using logically accurate and verifiable reasoning. By increasing the trust that customers can place in model responses, Automated Reasoning checks opens generative AI up to new use cases where accuracy is paramount.

Automated reasoning is a branch of AI that uses math to prove something is correct. It excels when dealing with problems where users need precise answers to a topic that is large and complex, and that has a well-defined set of rules or collection of knowledge about the subject. AWS has a team of world-class automated reasoning experts who have used this technology over the last decade to improve experiences across AWS, like proving that permissions and access controls are implemented correctly to enhance security or checking millions of scenarios across Amazon Simple Storage Service (S3) before deployment to ensure availability and durability remain protected.

Amazon Bedrock Guardrails makes it easy for customers to apply safety and responsible AI checks to generative AI applications, allowing customers to guide models to only talk about relevant topics. Accessible through Amazon Bedrock Guardrails, Automated Reasoning checks now allows Amazon Bedrock to validate factual responses for accuracy, produce auditable outputs, and show customers exactly why a model arrived at an outcome. This increases transparency and ensures that model responses are in line with a customer’s rules and policies.

Easily Build and Coordinate Multiple Agents to Execute Complex Workflows

As companies make generative AI a core part of their applications, they want to do more than just summarize content and power chat experiences. They also want their applications to take action. AI-powered agents can help customers’ applications accomplish these actions by using a model’s reasoning capabilities to break down a task, like helping with an order return or analyzing customer retention data, into a series of steps that the model can execute.

Amazon Bedrock Agents makes it easy for customers to build these agents to work across a company’s systems and data sources. While a single agent can be useful, more complex tasks, like performing financial analysis across hundreds or thousands of different variables, may require a large number of agents with their own specializations.

However, creating a system that can coordinate multiple agents, share context across them, and dynamically route different tasks to the right agent requires specialized tools and generative AI expertise that many companies do not have available. That is why AWS is expanding Amazon Bedrock Agents to support multi-agent collaboration, empowering customers to easily build and coordinate specialized agents to execute complex workflows.

Create Smaller, Faster, More Cost-Effective Models with Model Distillation

Customers today are experimenting with a wide variety of models to find the one best suited to the unique needs of their business. However, even with all the models available today, it is challenging to find one with the right mix of specific knowledge, cost, and latency. Larger models are more knowledgeable, but they take longer to respond and cost more, while small models are faster and cheaper to run, but are not as capable.

Model distillation is a technique that transfers the knowledge from a large model to a small model, while retaining the small model’s performance characteristics. However, doing this requires specialized machine learning (ML) expertise to work with training data, manually fine-tune the model, and adjust model weights without compromising the performance characteristics that led the customer to choose the smaller model in the first place.

With Amazon Bedrock Model Distillation, any customer can now distill their own model that can be up to 500% faster and 75% less expensive to run than original models, with less than 2% accuracy loss for use cases like retrieval augmented generation (RAG). Now, customers can optimize to achieve the best combination of capabilities, accuracy, latency, and cost for their use case—no ML expertise required.

With Amazon Bedrock Model Distillation, customers simply select the best model for a given use case and a smaller model from the same model family that delivers the latency their application requires at the right cost. After the customer provides sample prompts, Amazon Bedrock will do all the work to generate responses and fine-tune the smaller model, and it can even create more sample data, if needed, to complete the distillation process. This gives customers a model with the relevant knowledge and accuracy of the large model, but the speed and cost of the smaller model, making it ideal for production use cases, like real-time chat interactions. Model Distillation works with models from Anthropic, Meta, and the newly announced Amazon Nova Models.

Automated Reasoning checks, multi-agent collaboration, and Model Distillation are all available in preview.

About Amazon Web Services

Since 2006, Amazon Web Services has been the world’s most comprehensive and broadly adopted cloud. AWS has been continually expanding its services to support virtually any workload, and it now has more than 240 fully featured services for compute, storage, databases, networking, analytics, machine learning and artificial intelligence (AI), Internet of Things (IoT), mobile, security, hybrid, media, and application development, deployment, and management from 108 Availability Zones within 34 geographic regions, with announced plans for 18 more Availability Zones and six more AWS Regions in Mexico, New Zealand, the Kingdom of Saudi Arabia, Taiwan, Thailand, and the AWS European Sovereign Cloud. Millions of customers—including the fastest-growing startups, largest enterprises, and leading government agencies—trust AWS to power their infrastructure, become more agile, and lower costs. To learn more about AWS, visit aws.amazon.com.


Source: AWS

BigDATAwire