AWS Database Blog
Category: Customer Solutions
Real-time Iceberg ingestion with AWS DMS
Etleap is an AWS Advanced Technology Partner with the AWS Data & Analytics Competency and HAQM Redshift Service Ready designation. In this post, we show how Etleap helps you build scalable, near real-time pipelines that stream data from operational SQL databases into Iceberg tables using AWS DMS. You can use AWS DMS as a robust and configurable solution for change data capture (CDC) from all major databases into AWS.
Scaling HAQM RDS for MySQL performance for Careem’s digital platform on AWS
Careem powers rides, deliveries, and payments across the Middle East, North Africa and South Asia. As Careem grew, so did its data infrastructure challenges. Their monolithic 270 TB HAQM RDS for MySQL database consisting of one writer and five read replicas— experienced performance issues due to increased storage utilization, slow queries, high replica lag, and increased HAQM RDS cost. In this post, we provide a step-by-step breakdown of how Careem successfully implemented a phased data purging strategy, improving DB performance while addressing key technical challenges.
Zupee implements HAQM Neptune to detect Wallet transaction anomalies in real time
Zupee is a leading skill-based gaming platform offering casual and board games and is one of the fastest growing real money gaming platforms in India. Users can play multiple skill-based games online and win prizes. In this post, we show you how Zupee integrated HAQM Neptune Database to detect anomalies in real time for wallet transactions by creating a system for tracing the complex relationships between users, devices, and wallet transactions metadata.
How Habby enhanced resiliency and system robustness using Valkey GLIDE and HAQM ElastiCache
Habby is a game studio that creates interactive entertainment to connect players worldwide. We adopted Valkey GLIDE, a client library for HAQM ElastiCache for Valkey and Redis OSS, to address our system challenges. Our system uses the HAQM ElastiCache for Redis OSS publish/subscribe (Pub/Sub) functionality for the chat message sending. However, we faced challenges with connection stability during infrastructure changes, such as instance scaling, Redis OSS version upgrades, and hardware failures. This post describes our messaging system architecture and explains how we improved system reliability by using Valkey GLIDE as the client communicating with HAQM ElastiCache.
How HAQM Finance Automation built an operational data store with AWS purpose built databases to power critical finance applications
In this post, we discuss how the HAQM Finance Automation team used AWS purpose built databases, such as HAQM DynamoDB, HAQM OpenSearch Service, and HAQM Neptune together coupled with serverless compute like AWS Lambda to build an Operational Data Store (ODS) to store financial transactional data and support FinOps applications with millisecond latency. This data is the key enabler for FinOps business.
How Heroku migrated hundreds of thousands of self-managed PostgreSQL databases to HAQM Aurora
In this post, we discuss how Heroku migrated their multi-tenant PostgreSQL database fleet from self-managed PostgreSQL on HAQM Elastic Compute Cloud (HAQM EC2) to HAQM Aurora PostgreSQL-Compatible Edition. Heroku completed this migration with no customer impact, increasing platform reliability while simultaneously reducing operational burden. We dive into Heroku and their previous self-managed architecture, the new architecture, how the migration of hundreds of thousands of databases was performed, and the enhancements to the customer experience since its completion.
How Mindbody improved query latency and optimized costs using HAQM Aurora PostgreSQL Optimized Reads
In this post, we highlight the scaling and performance challenges Mindbody was facing due to an increase in their data growth. We also present the root cause analysis and recommendations for adopting to Aurora Optimized Reads, outlining the steps taken to address these issues. Finally, we discuss the benefits Mindbody realized from implementing these changes, including enhanced query performance, significant cost savings, and improved price predictability.
How GaadiBazaar reduced database costs by 40% with Aurora MySQL Serverless
GaadiBazaar draws on over 25 years of vehicle finance expertise from Cholamandalam to connect vehicle buyers and sellers. Their mission is to enable hassle-free transactions at fair prices through buyer-seller interactions and end-to-end financial assistance. This post shows you how GaadiBazaar, an online platform for buying and selling vehicles, achieved significant database cost savings by migrating to HAQM Aurora MySQL Compatible Edition Serverless.
How Aqua Security exports query data from HAQM Aurora to deliver value to their customers at scale
Aqua Security is the pioneer in securing containerized cloud native applications from development to production. Like many organizations, Aqua faced the challenge of efficiently exporting and analyzing large volumes of data to meet their business requirements. Specifically, Aqua needed to export and query data at scale to share with their customers for continuous monitoring and security analysis. In this post, we explore how Aqua addressed this challenge by using aws_s3.query_export_to_s3 function with their HAQM Aurora PostgreSQL-Compatible Edition and AWS Step Functions to streamline their query output export process, enabling scalable and cost-effective data analysis.
How Iterate.ai uses HAQM MemoryDB to accelerate and cost-optimize their workforce management conversational AI agent
Iterate.ai is an enterprise AI platform company delivering innovative AI solutions to industries such as retail, finance, healthcare, and quick-service restaurants. Among its standout offerings is Frontline, a workforce management platform powered by AI, designed to support and empower Frontline workers. Available on both the Apple App Store and Google Play, Frontline uses advanced AI tools to streamline operational efficiency and enhance communication among dispersed workforces. In this post, we give an overview of durable semantic caching in HAQM MemoryDB, and share how Iterate used this functionality to accelerate and cost-optimize Frontline.