article thumbnail

Managing Computer Vision Projects with Micha? Tadeusiak 

The MLOps Blog

Every episode is focused on one specific ML topic, and during this one, we talked to Michal Tadeusiak about managing computer vision projects. I’m joined by my co-host, Stephen, and with us today, we have Michal Tadeusiak , who will be answering questions about managing computer vision projects.

article thumbnail

Announcing Rekogniton Custom Moderation: Enhance accuracy of pre-trained Rekognition moderation models with your data

AWS Machine Learning Blog

Content moderation in Amazon Rekognition Amazon Rekognition is a managed artificial intelligence (AI) service that offers pre-trained and customizable computer vision capabilities to extract information and insights from images and videos. Upload images from your computer and provide labels. Choose Create project.

professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Optimize pet profiles for Purina’s Petfinder application using Amazon Rekognition Custom Labels and AWS Step Functions

AWS Machine Learning Blog

This post details how Purina used Amazon Rekognition Custom Labels , AWS Step Functions , and other AWS Services to create an ML model that detects the pet breed from an uploaded image and then uses the prediction to auto-populate the pet attributes. Start the model version when training is complete.

article thumbnail

Boost inference performance for Mixtral and Llama 2 models with new Amazon SageMaker containers

AWS Machine Learning Blog

This version offers support for new models (including Mixture of Experts), performance and usability improvements across inference backends, as well as new generation details for increased control and prediction explainability (such as reason for generation completion and token level log probabilities).

article thumbnail

Announcing provisioned concurrency for Amazon SageMaker Serverless Inference

AWS Machine Learning Blog

In addition, you can now use Application Auto Scaling with provisioned concurrency to address inference traffic dynamically based on target metrics or a schedule. In this post, we discuss what provisioned concurrency and Application Auto Scaling are, how to use them, and some best practices and guidance for your inference workloads.

article thumbnail

Improve performance of Falcon models with Amazon SageMaker

AWS Machine Learning Blog

The decode phase includes the following: Completion – After the prefill phase, you have a partially generated text that may be incomplete or cut off at some point. The decode phase is responsible for completing the text to make it coherent and grammatically correct. The default is 32.

article thumbnail

Deploy Falcon-40B with large model inference DLCs on Amazon SageMaker

AWS Machine Learning Blog

LMI DLCs are a complete end-to-end solution for hosting LLMs like Falcon-40B. You can monitor the status of the endpoint by calling DescribeEndpoint , which will tell you when everything is complete. His expertise lies in Deep Learning in the domains of Natural Language Processing (NLP) and Computer Vision.