Exclusive Content:

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

Running Your ML Notebook on Databricks: A Step-by-Step Guide

A Step-by-Step Guide to Hosting Machine Learning Notebooks in...

“Revealing Weak Infosec Practices that Open the Door for Cyber Criminals in Your Organization” • The Register

Warning: Stolen ChatGPT Credentials a Hot Commodity on the...

Scaling Global Customer Support: How Ring Utilizes Amazon Bedrock Knowledge Bases

Building a Scalable Multi-Locale Support Chatbot at Ring Using Amazon Bedrock


Streamlining Global Support Through Retrieval-Augmented Generation

This post, cowritten with David Kim and Premjit Singh from Ring, explores how Ring developed a robust multi-locale support chatbot that enhances customer experiences while optimizing costs.

The Support Evolution Journey for Ring

Requirements for a RAG-Based Support System

Global Content Localization

Serverless, Managed Architecture

Scalable Knowledge Management

Performance and Cost Optimization

Overview of Solution

Ingestion and Evaluation Workflow

Promotion Workflow: Customer-Facing

Other Considerations for Your Implementation

Conclusion

About the Authors


This heading structure effectively organizes the content while highlighting the main topics covered in the post.

Scaling Self-Service Support Globally: Ring’s Journey with RAG-Based Chatbots

This post is cowritten with David Kim and Premjit Singh from Ring.

In an age where customer service expectations are higher than ever, scaling self-service support globally comes with its own set of challenges. Beyond mere translation, it requires a robust system capable of localizing content and maintaining a seamless customer experience across different regions. In this post, we explore how Ring, Amazon’s home security subsidiary, successfully built a production-ready, multi-locale Retrieval-Augmented Generation (RAG)-based support chatbot using Amazon Bedrock Knowledge Bases.

The Need for a Transformation

Initially, Ring relied on a rule-based chatbot built with Amazon Lex. While functional, this system fell short in terms of adaptability, often resulting in a 16% escalation rate to human agents during peak periods. As Ring expanded into international locales, the limitations of this system became clear. The team needed a solution that not only provided accurate support but also did so cost-effectively across multiple regions.

Four Requirements for a RAG-Based Support System

To overcome these challenges, Ring identified four core requirements that guided their architectural approach:

  1. Global Content Localization: Ring’s international presence required more than just translation. It demanded region-specific product information, taking into account factors like voltage specifications and regulatory compliance.

  2. Serverless, Managed Architecture: To enable their engineering team to concentrate on improving the customer experience, Ring sought a fully-managed, serverless solution.

  3. Scalable Knowledge Management: With hundreds of documents to manage, Ring required a sophisticated vector search technology capable of retrieving precise information from a unified repository.

  4. Performance and Cost Optimization: The architecture needed to meet stringent latency requirements while minimizing operational complexities and costs. Cross-region latency accounted for a mere 10% of total response time, which allowed for a centralized approach to infrastructure.

Designing the Solution

Ring’s architecture seamlessly integrates various AWS services like Amazon Bedrock Knowledge Bases, AWS Lambda, AWS Step Functions, and Amazon S3, effectively addressing their requirements.

Ingestion and Evaluation Workflow

The Ring content team uploads support documentation to Amazon S3. This structured upload not only saves raw data but also extracts metadata for effective knowledge retrieval. The proposed architecture orchestrates a daily knowledge base creation and evaluation process using AWS Step Functions, thus facilitating independent evaluation and easy rollbacks.

Promotion Workflow: Customer-Facing Interactions

When customers interact with the chatbot, their queries are processed through Lambda functions that apply metadata filtering for precise regional targeting. This ensures that customers receive relevant and context-specific information.

Operational Best Practices

When building your own RAG-based system at scale, several architectural approaches and operational strategies should be considered:

Vector Store Selection

Choosing the right vector store is crucial. While the Ring implementation uses Amazon OpenSearch Serverless for advanced search capabilities, alternative options like Amazon S3 vectors also provide automatic scaling and durability at a lower cost.

Versioning Architecture Considerations

Implementing a versioning strategy—such as maintaining separate Knowledge Bases for each version—can simplify rollback capabilities. However, this requires management of multiple instances, which can add complexity.

Disaster Recovery: Multi-Region Deployment

To ensure robust disaster recovery, it’s essential to deploy your architecture across multiple regions. This includes creating Knowledge Base instances, maintaining cross-region copies, and implementing data synchronization processes.

Foundation Model Throughput: Cross-Region Inference

Cross-Region inference (CRIS) allows for scaling model throughput during traffic bursts, ensuring your system can handle increased demand without faltering.

Embedding Model Selection and Chunking Strategy

The choice of embedding models and chunking strategies can significantly influence retrieval accuracy and response quality. For instance, Ring uses the Amazon Titan Embeddings model, which proved effective for their documentation.

Conclusion

By implementing a sophisticated, multi-locale RAG-based support chatbot architecture, Ring achieved a 21% reduction in operational costs while providing consistent customer experiences across ten international regions. This achievement not only highlights the power of leveraging AWS services like Amazon Bedrock Knowledge Bases but also showcases the potential for other organizations to adopt similar strategies for optimizing their customer support systems.

As Ring continues to evolve its chatbot architecture, integrating specialized agents for specific tasks, the future looks promising for RAG-based systems built on AWS. For organizations looking to harness the capabilities of automated, localized support, exploring these architectural patterns could drive significant innovation and efficiency.

To learn more about Amazon Bedrock Knowledge Bases, visit the Amazon Bedrock documentation.

About the Authors

Gopinath Jagadesan

Senior Solution Architect at AWS passionate about generative AI and real-world applications.

David Kim

Software Development Engineer at Ring, focusing on conversational AI and multi-agent systems.

Premjit Singh

Software Development Manager at Ring, dedicated to enhancing customer experiences through AWS AI services.


By sharing insights from their journey, the authors hope to inspire others in the tech community to rethink their approach to customer support solutions. With the right tools and architecture, efficient and impactful self-service support is within reach.

Latest

Are AI Chatbots Creating the Next Walled Garden?

The Rising Tide of AI Chatbots: Balancing Convenience with...

UK Advances in Quantum Communications with New Space Mission Launch

Pioneering UK Mission Launches Satellite Platform for Optical Quantum...

I Turned to ChatGPT for Solutions to 7 Daily Challenges — The Results Surprised Me!

Transforming Overwhelm: How ChatGPT Helped Me Tackle Everyday Stressors 1....

European Student Robotics Association Links University Robotics Talent Across Borders — EdTech Innovation Hub

Uniting European Robotics Talent: The Initiative of the European...

Don't miss

Haiper steps out of stealth mode, secures $13.8 million seed funding for video-generative AI

Haiper Emerges from Stealth Mode with $13.8 Million Seed...

Running Your ML Notebook on Databricks: A Step-by-Step Guide

A Step-by-Step Guide to Hosting Machine Learning Notebooks in...

VOXI UK Launches First AI Chatbot to Support Customers

VOXI Launches AI Chatbot to Revolutionize Customer Services in...

Investing in digital infrastructure key to realizing generative AI’s potential for driving economic growth | articles

Challenges Hindering the Widescale Deployment of Generative AI: Legal,...

Addressing LLM Hallucinations in Regulated Industries: The Deterministic Models of Artificial...

Navigating the Paradox of Large Language Models in Regulated Industries Addressing Hallucinations for Safe AI Adoption in Financial and Healthcare Sectors The Evolution of Language Models:...

A Technical Guide to Reinforcement Fine-Tuning on Amazon Bedrock Using OpenAI-Compatible...

Unlocking Powerful Customization: Reinforcement Fine-Tuning on Amazon Bedrock A Comprehensive Guide to Implementing RFT for Enhanced Model Performance How Reinforcement Fine-Tuning Works The Core Concept: Learning from...

Enhancing LLM Fine-Tuning with Unstructured Data through SageMaker Unified Studio and...

Integrating Amazon SageMaker Unified Studio with S3 for Enhanced Visual Question Answering Using Llama 3.2 Overview of the Integration Last year, AWS announced an integration between...