Amazon Web Services Hit By Slowdown - InformationWeek

InformationWeek is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

Cloud // Infrastructure as a Service
03:48 PM
Connect Directly

Amazon Web Services Hit By Slowdown

AWS' HMSCloud encryption key storage service produced network errors in part of Zone B, causing system delays and failovers.

8 Great Cloud Storage Services
8 Great Cloud Storage Services
(click image for larger view and for slideshow)
An encryption key storage service inside Amazon Web Services, the Hardware Security Module appliance, was affected by connectivity problems over a period of an hour and 18 minutes in one availability zone Friday morning.

The slowdown affected only one availability zone, according to the Amazon Service Health Dashboard, but it was severe enough to trigger a failover to alternative systems for latency-sensitive customers. One customer,, an API management firm in Spain, found its production systems failing over to other availability zones in US-East and US-West, as planned, in the event of business-risking operational problems.

The HSMCloud service is designed to allow customers to store encryption keys in a secure manner inside AWS, keeping the keys close to the customer's virtual machines running in an Amazon Virtual Private Cloud. Without the HSM, the keys need to be kept on-premises and retrieved each time an encryption function is performed, slowing operations. US-East is Amazon's most heavily used data center complex with five availability zones, located in Ashburn, Va.

AWS Service Health Dashboard reported that HSMCloud service returned to normal at 9:22 a.m. Pacific. As usual, the company offered no explanation beyond the cryptic notices posted to the dashboard.

[Want more on Amazon's occasional outages? See Amazon's Dec. 24 Outage: A Closer Look. ]

In addition to HSMCloud, other services in the Northern Virginia facility were affected by connectivity problems at about the same time. They included EC2 compute, Simple Email Service, Elastic Load Balancing, Relational Database Service and the RedShift data warehousing service. In each case, a single availability zone was affected.

Affected customers said the connectivity problems appeared to arise in the B availability zone. Each availability zone appears to be made up of several discrete data center sections.

Rhommel Lamas, a systems operations engineer for, said its production system in one section of the B availability suffered major latencies as it tried to connect to D and E availability zones. It also suffered the same latency as it attempted to connect to another B zone section, according to a map drawn by's monitoring service, Boundary. The latencies amounted to 1,260ms or 1.26 seconds, a crippling delay for the latency-sensitive business of managing customers' APIs. Such a latency would back up thousands of API requests on highly trafficked websites, putting the customers of at risk of losing site visitors and business.

Lamas in a telephone interview said he spotted the delays building in his Boundary monitoring system a few minutes before Amazon reported them on its dashboard. production systems automatically fail over to backup systems when latencies reach a certain threshold.

"We have a complex architecture and this is just one tiny part of it," Lamas said in an email message preceding the telephone interview. "We saw how all of our Region B on US-East was failing with increasing latency issues and errors between machines in different zones," he wrote. lost messaging packets as well in attempted communications between the zones.

Lamas said his firm didn't suffer any loss of business because automated failovers worked as planned. "No, it didn't affect our ability to serve our customers," he said. It wasn't immediately clear whether Amazon SLAs that offer replacement time for any lost in a service outage would offset's increased instance cost.

Other Amazon users noticed the problem. AWS said it began at 7:32 a.m. Pacific. At 8:34 a.m., Joshua Frattarola (@jfrattarola) tweeted: "Friday 13th starts off with major network issues affecting all of AWS US-EAST-1 Region. Probably somebody typing Google into Google."

We welcome your comments on this topic on our social media channels, or [contact us directly] with questions about the site.
Comment  | 
Print  | 
More Insights
InformationWeek Is Getting an Upgrade!

Find out more about our plans to improve the look, functionality, and performance of the InformationWeek site in the coming months.

Pandemic Responses Make Room for More Data Opportunities
Jessica Davis, Senior Editor, Enterprise Apps,  5/4/2021
10 Things Your Artificial Intelligence Initiative Needs to Succeed
Lisa Morgan, Freelance Writer,  4/20/2021
Transformation, Disruption, and Gender Diversity in Tech
Joao-Pierre S. Ruth, Senior Writer,  5/6/2021
White Papers
Register for InformationWeek Newsletters
Current Issue
Planning Your Digital Transformation Roadmap
Download this report to learn about the latest technologies and best practices or ensuring a successful transition from outdated business transformation tactics.
Flash Poll