Remove Blog Remove Data Ingestion Remove DevOps Remove ML
article thumbnail

How Axfood enables accelerated machine learning throughout the organization using Amazon SageMaker

AWS Machine Learning Blog

In this post, we share how Axfood, a large Swedish food retailer, improved operations and scalability of their existing artificial intelligence (AI) and machine learning (ML) operations by prototyping in close collaboration with AWS experts and using Amazon SageMaker. This is a guest post written by Axfood AB.

article thumbnail

Deliver your first ML use case in 8–12 weeks

AWS Machine Learning Blog

Do you need help to move your organization’s Machine Learning (ML) journey from pilot to production? Most executives think ML can apply to any business decision, but on average only half of the ML projects make it to production. Challenges Customers may face several challenges when implementing machine learning (ML) solutions.

ML 88
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Boost employee productivity with automated meeting summaries using Amazon Transcribe, Amazon SageMaker, and LLMs from Hugging Face

AWS Machine Learning Blog

The service allows for simple audio data ingestion, easy-to-read transcript creation, and accuracy improvement through custom vocabularies. Hugging Face is an open-source machine learning (ML) platform that provides tools and resources for the development of AI projects.

article thumbnail

How Earth.com and Provectus implemented their MLOps Infrastructure with Amazon SageMaker

AWS Machine Learning Blog

This blog post is co-written with Marat Adayev and Dmitrii Evstiukhin from Provectus. When machine learning (ML) models are deployed into production and employed to drive business decisions, the challenge often lies in the operation and management of multiple models. The models developed by Earth.com lived across various notebooks.

DevOps 94
article thumbnail

Foundational models at the edge

IBM Journey to AI blog

Foundational models (FMs) are marking the beginning of a new era in machine learning (ML) and artificial intelligence (AI) , which is leading to faster development of AI that can be adapted to a wide range of downstream tasks and fine-tuned for an array of applications. Large language models (LLMs) have taken the field of AI by storm.

article thumbnail

Introducing the Amazon Comprehend flywheel for MLOps

AWS Machine Learning Blog

This combination of great models and continuous adaptation is what will lead to a successful machine learning (ML) strategy. MLOps focuses on the intersection of data science and data engineering in combination with existing DevOps practices to streamline model delivery across the ML development lifecycle.

article thumbnail

End-to-End Pipeline for Segmentation with TFX, Google Cloud, and Hugging Face

TensorFlow

Posted by Chansung Park, Sayak Paul (ML and Cloud GDEs) TensorFlow Extended ( TFX ) is a flexible framework allowing Machine Learning (ML) practitioners to iterate on production-grade ML workflows faster with reliability and resiliency. When it comes to ML projects, these changes usually affect the model and/or data.