Exclusive Content:

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

“Revealing Weak Infosec Practices that Open the Door for Cyber Criminals in Your Organization” • The Register

Warning: Stolen ChatGPT Credentials a Hot Commodity on the...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Sprinklr boosts efficiency by 20% and cuts expenses by 25% for machine learning inference on AWS Graviton3

Optimizing AI Workloads with AWS Graviton3: A Case Study with Sprinklr

In today’s digital age, artificial intelligence (AI) has become an indispensable tool for businesses looking to streamline processes, gather insights, and improve decision-making. With the vast amount of data generated daily, companies need AI solutions that can handle diverse workloads efficiently while also being cost-effective. Sprinklr, a leading provider of customer experience management software, has achieved remarkable success in optimizing mixed AI workload inference performance with the help of AWS Graviton3-based instances.

Sprinklr’s AI processes thousands of servers to fine-tune and serve over 750 pre-built AI models across various verticals, making more than 10 billion predictions per day. To deliver tailored user experiences, Sprinklr deploys specialized AI models for specific business applications that utilize nine layers of machine learning, extracting meaning from data across different formats. This diverse database of models presents unique challenges for choosing the most efficient deployment infrastructure that balances scale and cost-effectiveness.

For mixed AI workloads with real-time latency requirements, traditional production instances were not cost-effective due to smaller model sizes and infrequent inference requests. This led Sprinklr to explore new instances that could offer the right balance of performance and cost-efficiency. AWS Graviton3 processors emerged as the ideal choice, offering optimized support for ML workloads and significant improvements in performance compared to x86-based instances.

By migrating mixed inference/search workloads to Graviton3-based c7g instances, Sprinklr achieved impressive results:
– 20% throughput improvement and 30% latency reduction
– 25-30% cost savings
– Improved customer experience through reduced latency and increased throughput
– Lower carbon footprint by running more efficiently on the same number of instances

The transition to Graviton3-based instances was seamless, with engineering time kept minimal and performance improvements quickly realized in production workloads. Sprinklr has already migrated several models to Graviton3-based instances and plans to move more models to further optimize performance and cost savings.

In conclusion, Sprinklr’s success with AWS Graviton3-based instances serves as a testament to the importance of adopting cutting-edge technologies for efficiency, cost-saving, and improved customer experience. As technology continues to evolve, companies must stay ahead of the curve and seek new compute solutions that not only reduce costs but also enhance their products and services.

Latest

Deploy Geospatial Agents Using Foursquare Spatial H3 Hub and Amazon SageMaker AI

Transforming Geospatial Analysis: Deploying AI Agents for Rapid Spatial...

ChatGPT Transforms into a Full-Fledged Chat App

ChatGPT Introduces Group Chat Feature: Prove Your Point with...

Sunday Bucks Introduces Mainstream Training Techniques for Teaching Robots to Load Dishes

Sunday Robotics Unveils Memo: A Revolutionary Autonomous Home Robot Transforming...

Ubisoft Unveils Playable Generative AI Experiment

Ubisoft Unveils 'Teammates': A Generative AI-R Powered NPC Experience...

Don't miss

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Investing in digital infrastructure key to realizing generative AI’s potential for driving economic growth | articles

Challenges Hindering the Widescale Deployment of Generative AI: Legal,...

Microsoft launches new AI tool to assist finance teams with generative tasks

Microsoft Launches AI Copilot for Finance Teams in Microsoft...

Optimize AI Operations with the Multi-Provider Generative AI Gateway Architecture

Streamlining AI Management with the Multi-Provider Generative AI Gateway on AWS Introduction to the Generative AI Gateway Addressing the Challenge of Multi-Provider AI Infrastructure Reference Architecture for...

MSD Investigates How Generative AI and AWS Services Can Enhance Deviation...

Transforming Deviation Management in Biopharmaceuticals: Harnessing Generative AI and Emerging Technologies at MSD Transforming Deviation Management in Biopharmaceutical Manufacturing with Generative AI Co-written by Hossein Salami...

Best Practices and Deployment Patterns for Claude Code Using Amazon Bedrock

Deploying Claude Code with Amazon Bedrock: A Comprehensive Guide for Enterprises Unlock the power of AI-driven coding assistance with this step-by-step guide to deploying Claude...