AWS Database Blog
Zupee implements Amazon Neptune to detect Wallet transaction anomalies in real time
Zupee is a leading skill-based gaming platform offering casual and board games and is one of the fastest growing real money gaming platforms in India. Users can play multiple skill-based games online and win prizes. In this post, we show you how Zupee integrated Amazon Neptune Database to detect anomalies in real time for wallet transactions by creating a system for tracing the complex relationships between users, devices, and wallet transactions metadata.
How Habby enhanced resiliency and system robustness using Valkey GLIDE and Amazon ElastiCache
Habby is a game studio that creates interactive entertainment to connect players worldwide. We adopted Valkey GLIDE, a client library for Amazon ElastiCache for Valkey and Redis OSS, to address our system challenges. Our system uses the Amazon ElastiCache for Redis OSS publish/subscribe (Pub/Sub) functionality for the chat message sending. However, we faced challenges with connection stability during infrastructure changes, such as instance scaling, Redis OSS version upgrades, and hardware failures. This post describes our messaging system architecture and explains how we improved system reliability by using Valkey GLIDE as the client communicating with Amazon ElastiCache.
Migrate SQL Server user databases from Amazon EC2 to Amazon RDS Custom using Amazon EBS snapshots
In this post, we present a practical approach to one of the most significant challenges organizations face when adopting Amazon RDS Custom for SQL Server: migrating large datasets from SQL Server on Amazon EC2 to Amazon RDS Custom for SQL Server efficiently and cost-effectively. By using SQL Server’s native detach and attach method combined with EBS snapshots, you can migrate your databases without requiring Amazon S3 or AWS DMS.
Choose the right throughput strategy for Amazon DynamoDB applications
When getting started with DynamoDB, one of the first decisions you will make is choosing between two throughput modes: on-demand and provisioned. On-demand mode is the default and recommended throughput option because it simplifies building modern, serverless applications that can start small and scale to millions of requests per second. However, choosing the right throughput strategy requires evaluating your operational needs, development velocity, and application characteristics, with cost being a key consideration. In this post, we examine both throughput modes in detail, exploring their characteristics, strengths, and ideal use cases.
Best practices to handle AWS DMS tasks during PostgreSQL upgrades
When you decide to upgrade your PostgreSQL database which is configured as source or target for an ongoing AWS DMS task, it’s important to factor this into your upgrade planning. In this post, we discuss the best practices to handle the AWS DMS tasks during PostgreSQL upgrades to minor or major versions.
Integrate your Spring Boot application with Amazon ElastiCache
In this post, we explore the basics of integrating a Spring Boot application with ElastiCache to enable caching. Amazon ElastiCache is a fully managed, Valkey-, Memcached-, and Redis OSS-compatible service that delivers real-time, cost-optimized performance for modern applications with 99.99% SLA availability. ElastiCache speeds up application performance, scaling to millions of operations per second with microsecond response time.
How Amazon Finance Automation built an operational data store with AWS purpose built databases to power critical finance applications
In this post, we discuss how the Amazon Finance Automation team used AWS purpose built databases, such as Amazon DynamoDB, Amazon OpenSearch Service, and Amazon Neptune together coupled with serverless compute like AWS Lambda to build an Operational Data Store (ODS) to store financial transactional data and support FinOps applications with millisecond latency. This data is the key enabler for FinOps business.
Challenges and strategies of migrating a high-throughput relational database
In this post, we explore key strategies and AWS tools to help you successfully migrate your high-throughput relational database while minimizing business disruption.
How Heroku migrated hundreds of thousands of self-managed PostgreSQL databases to Amazon Aurora
In this post, we discuss how Heroku migrated their multi-tenant PostgreSQL database fleet from self-managed PostgreSQL on Amazon Elastic Compute Cloud (Amazon EC2) to Amazon Aurora PostgreSQL-Compatible Edition. Heroku completed this migration with no customer impact, increasing platform reliability while simultaneously reducing operational burden. We dive into Heroku and their previous self-managed architecture, the new architecture, how the migration of hundreds of thousands of databases was performed, and the enhancements to the customer experience since its completion.
Using generative AI and Amazon Bedrock to generate SPARQL queries to discover protein functional information with UniProtKB and Amazon Neptune
In this post, we demonstrate how to use generative AI and Amazon Bedrock to transform natural language questions into graph queries to run against a knowledge graph. We explore the generation of queries written in the SPARQL query language, a well-known language for querying a graph whose data is represented as Resource Description Framework (RDF).