Remove LLM Remove Metadata Remove Software Development
article thumbnail

Reducing hallucinations in LLM agents with a verified semantic cache using Amazon Bedrock Knowledge Bases

AWS Machine Learning Blog

Similar to how a customer service team maintains a bank of carefully crafted answers to frequently asked questions (FAQs), our solution first checks if a users question matches curated and verified responses before letting the LLM generate a new answer. No LLM invocation needed, response in less than 1 second.

LLM 119
article thumbnail

Time series forecasting with LLM-based foundation models and scalable AIOps on AWS

AWS Machine Learning Blog

However, traditional machine learning approaches often require extensive data-specific tuning and model customization, resulting in lengthy and resource-heavy development. Enter Chronos , a cutting-edge family of time series models that uses the power of large language model ( LLM ) architectures to break through these hurdles.

LLM 109
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Build agentic systems with CrewAI and Amazon Bedrock

Flipboard

Consider a software development use case AI agents can generate, evaluate, and improve code, shifting software engineers focus from routine coding to more complex design challenges. Agentic systems, on the other hand, are designed to bridge this gap by combining the flexibility of context-aware systems with domain knowledge.

LLM 174
article thumbnail

Evaluate large language models for your machine translation tasks on AWS

AWS Machine Learning Blog

However, the industry is seeing enough potential to consider LLMs as a valuable option. The following are a few potential benefits: Improved accuracy and consistency LLMs can benefit from the high-quality translations stored in TMs, which can help improve the overall accuracy and consistency of the translations produced by the LLM.

article thumbnail

Build a dynamic, role-based AI agent using Amazon Bedrock inline agents

AWS Machine Learning Blog

For this demo, weve implemented metadata filtering to retrieve only the appropriate level of documents based on the users access level, further enhancing efficiency and security. The role information is also used to configure metadata filtering in the knowledge bases to generate relevant responses.

article thumbnail

Transforming financial analysis with CreditAI on Amazon Bedrock: Octus’s journey with AWS

AWS Machine Learning Blog

With this LLM, CreditAI was now able to respond better to broader, industry-wide queries than before. It also enables economies of scale with development velocity given that over 75 engineers at Octus already use AWS services for application development.

DevOps 87
article thumbnail

Ground truth generation and review best practices for evaluating generative AI question-answering with FMEval

AWS Machine Learning Blog

To scale ground truth generation and curation, you can apply a risk-based approach in conjunction with a prompt-based strategy using LLMs. Its important to note that LLM-generated ground truth isnt a substitute for use case SME involvement. To convert the source document excerpt into ground truth, we provide a base LLM prompt template.