Remove Conversational AI Remove LLM Remove Responsible AI
article thumbnail

Advancing AI trust with new responsible AI tools, capabilities, and resources

AWS Machine Learning Blog

As generative AI continues to drive innovation across industries and our daily lives, the need for responsible AI has become increasingly important. At AWS, we believe the long-term success of AI depends on the ability to inspire trust among users, customers, and society.

article thumbnail

Evaluate conversational AI agents with Amazon Bedrock

AWS Machine Learning Blog

However, the dynamic and conversational nature of these interactions makes traditional testing and evaluation methods challenging. Conversational AI agents also encompass multiple layers, from Retrieval Augmented Generation (RAG) to function-calling mechanisms that interact with external knowledge sources and tools.

professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Enhance conversational AI with advanced routing techniques with Amazon Bedrock

AWS Machine Learning Blog

Amazon Bedrock is a fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies like AI21 Labs, Anthropic, Cohere, Meta, Stability AI, and Amazon using a single API, along with a broad set of capabilities you need to build generative AI applications with security, privacy, and responsible AI.

article thumbnail

Improve factual consistency with LLM Debates

AWS Machine Learning Blog

In this post, we demonstrate the potential of large language model (LLM) debates using a supervised dataset with ground truth. In this LLM debate, we have two debater LLMs, each one taking one side of an argument and defending it based on the previous arguments for N(=3) rounds. The arguments are saved for a judge LLM to review.

LLM 93
article thumbnail

Improve LLM application robustness with Amazon Bedrock Guardrails and Amazon Bedrock Agents

AWS Machine Learning Blog

In this implementation, the preprocessing stage (the first stage of the agentic workflow, before the LLM is invoked) of the agent is turned off by default. To learn more about using agents to orchestrate workflows, see Automate tasks in your application using conversational agents.

LLM 127
article thumbnail

The GenAI Frontier: 10 Transformative LLM Research Papers of 2023 from LLaMA to GPT-4

Topbots

Top LLM Research Papers 2023 1. LLaMA by Meta AI Summary The Meta AI team asserts that smaller models trained on more tokens are easier to retrain and fine-tune for specific product applications. The instruction tuning involves fine-tuning the Q-Former while keeping the image encoder and LLM frozen.

LLM 106
article thumbnail

Build safe and responsible generative AI applications with guardrails

AWS Machine Learning Blog

However, the implementation of LLMs without proper caution can lead to the dissemination of misinformation , manipulation of individuals, and the generation of undesirable outputs such as harmful slurs or biased content. Introduction to guardrails for LLMs The following figure shows an example of a dialogue between a user and an LLM.