On Oct 22, 2025, a subset of customers in us-east-1 experienced publishing delays due to a service degradation in our event processing infrastructure.
The time required for content actions to complete processing varied by customer activity and overall publishing volume.
During this period, customers may have observed delays in:
On Oct 22, 2025, our monitoring infrastructure detected elevated system usage, which led to delays in publishing content. Shortly after, we identified that one of our core event processing clusters had reached near-full storage capacity too quickly, which caused a slowdown in content updates.
As teams worked to expand capacity, the system experienced additional delays due to network connectivity issues within one of the cluster nodes. This created a backlog of publishing events, resulting in slower-than-normal delivery of updates to customer sites.
All times ET + 24 hour clock
| Time | Event |
|---|---|
| Oct 22, 2025 | |
| 16:23 | Initial alert received regarding cluster performance. |
| 16:30 – 17:30 | Increased system storage to address capacity constraints. |
| 17:30–18:30 | Investigated and resolved AWS network connectivity issues affecting message delivery. |
| Oct 23, 2025 | |
| 00:00 - 06:00 | Continued scaling efforts and cleared high-priority publishing events. |
| 05:00 | Identified and resolved a secondary issue related to message distribution within the cluster. |
| 06:00 - 10:00 | Processed remaining low-priority updates to stabilize system throughput. All lower-priority queues were temporarily escalated to ensure full recovery. |
| 13:17 | All systems verified stable; incident resolved and “all clear” issued. |