Optimizing Amazon Easy Queue Service (SQS) for pace and scale


Voiced by Polly

After a number of public betas, we launched Amazon Easy Queue Service (Amazon SQS) in 2006. Practically twenty years later, this totally managed service continues to be a basic constructing block for microservices, distributed techniques, and serverless functions, processing over 100 million messages per second at peak instances.

As a result of there’s at all times a greater approach, we proceed to search for methods to enhance efficiency, safety, inside effectivity, and so forth. Once we do discover a potential solution to do one thing higher, we’re cautious to protect present habits, and sometimes run new and outdated techniques in parallel to permit us to match outcomes.

Immediately I wish to let you know how we lately made enhancements to Amazon SQS to cut back latency, improve fleet capability, mitigate an approaching scalability cliff, and cut back energy consumption.

Enhancing SQS
Like many AWS providers, Amazon SQS is applied utilizing a group of inside microservices. Let’s deal with two of them immediately:

Buyer Entrance-Finish – The client-facing front-end accepts, authenticates, and authorizes API calls corresponding to CreateQueue and SendMessage. It then routes every request to the storage back-end.

Storage Again-Finish -This inside microservice is accountable for persisting messages despatched to plain (non-FIFO) queues. Utilizing a cell-based mannequin, every cluster within the cell incorporates a number of hosts, every buyer queue is assigned to a number of clusters, and every cluster is accountable for a large number of queues:

Connections – Previous and New
The unique implementation used a connection per request between these two providers. Every front-end had to hook up with many hosts, which mandated the usage of a connection pool, and in addition risked reaching an final, hard-wired restrict on the variety of open connections. Whereas it’s usually potential to easily throw {hardware} at issues like this and scale out, that’s not at all times one of the simplest ways. It merely strikes the second of reality (the “scalability cliff”) into the long run and doesn’t make environment friendly use of assets.

After fastidiously contemplating a number of long-term options, the Amazon SQS group invented a brand new, proprietary binary framing protocol between the shopper front-end and storage back-end. The protocol multiplexes a number of requests and responses throughout a single connection, utilizing 128-bit IDs and checksumming to stop crosstalk. Server-side encryption gives an extra layer of safety towards unauthorized entry to queue knowledge.

It Works!
The brand new protocol was put into manufacturing earlier this yr and has processed 744.9 trillion requests as I write this. The scalability cliff has been eradicated and we’re already on the lookout for methods to place this new protocol to work in different methods.

Efficiency-wise, the brand new protocol has lowered dataplane latency by 11% on common, and by 17.4% on the P90 mark. Along with making SQS itself extra performant, this alteration advantages providers that construct on SQS as effectively. For instance, messages despatched via Amazon Easy Notification Service (Amazon SNS) now spend 10% much less time “inside” earlier than being delivered. Lastly, as a result of protocol change, the prevailing fleet of SQS hosts (a mixture of X86 and Graviton-powered situations) can now deal with 17.8% extra requests than earlier than.

Extra to Come
I hope that you’ve loved this little peek contained in the implementation of Amazon SQS. Let me know within the feedback, and I’ll see if I can discover some extra tales to share.

Jeff;



Recent Articles

Related Stories

Leave A Reply

Please enter your comment!
Please enter your name here

Stay on op - Ge the daily news in your inbox