Update - We are continuing to investigate this issue.
Apr 20, 2026 - 15:11 UTC
Update - We have not received any further updates from AWS at this time. However, we are actively monitoring the outage and will provide additional information as it becomes available. Also, please continue to refer to the AWS status page for more detailed updates. https://health.aws.amazon.com/health/status
All the guidance previously included about stack migration is still relevant. Please reach out to our Support team if you have any questions.
Mar 19, 2026 - 12:13 UTC
Update - We are actively monitoring the situation, but at this time there are no new updates to share. The next update will be provided once we have more information to share. Please reach out to our Support team if you have any questions.
Mar 04, 2026 - 22:22 UTC
Update - We are continuing to investigate this issue.
Mar 04, 2026 - 10:28 UTC
We will provide updates when we have them, but we do not have an expected resolution time at this point.
Mar 02, 2026 - 10:31 UTC
Update - Customers are recommended to configure a new blank stack in an alternative Grafana Cloud region and to reconfigure their clients (such as Grafana Alloy) to send telemetry to that region, Fleet Management can be used for this purpose https://grafana.com/docs/grafana-cloud/send-data/fleet-management/introduction/
Mar 02, 2026 - 10:04 UTC
Update - We are updating this incident to reflect a complete outage in prod-me-central-1, due to an on-going AWS UAE data center issue. We will provide further updates accordingly.
Mar 02, 2026 - 08:36 UTC
Update - We are observing write and read outage errors across all databases (metrics, logs, traces) in prod-me-central-1, due to an on-going AWS UAE data center issue. We will provide further updates accordingly.
Mar 02, 2026 - 08:21 UTC
Update - We are observing write and read outage errors across all databases (metrics, logs, traces) in prod-me-central-1, due to an on-going AWS UAE data center issue. We will provide further updates accordingly.
Mar 02, 2026 - 08:14 UTC
Investigating - We are seeing elevated write and read path errors in prod-me-central-1, due to an on-going AWS UAE data center issue. We will provide further updates accordingly.
Mar 02, 2026 - 06:43 UTC
Grafana Cloud: Hosted Grafana
Major Outage
AWS Australia - prod-ap-southeast-2
Operational
AWS Brazil - prod-sa-east-1
Operational
AWS Canada - prod-ca-east-0
Operational
AWS Germany - prod-eu-west-2
Operational
AWS Germany - prod-eu-west-4
Operational
AWS India - prod-ap-south-1
Operational
AWS Japan - prod-ap-northeast-0
Operational
AWS UAE - prod-me-central-1
Major Outage
AWS Singapore - prod-ap-southeast-1
Operational
AWS Sweden - prod-eu-north-0
Operational
AWS US East - prod-us-east-0
Operational
AWS US East - prod-us-east-2
Operational
AWS US West - prod-us-west-0
Operational
AWS Australia - prod-au-southeast-1
Operational
AWS UK - prod-gb-south-1
Operational
AWS Ireland - prod-eu-west-6
Operational
Azure US Central - us-central2
Operational
AWS Switzerland - prod-eu-central-0
Operational
Azure Netherlands - prod-eu-west-3
Operational
GCP Australia - prod-au-southeast-0
Operational
GCP Belgium - prod-eu-west-0
Operational
GCP Brazil - prod-sa-east-0
Operational
GCP India - prod-ap-south-0
Operational
GCP Singapore - prod-ap-southeast-0
Operational
GCP UK - prod-gb-south-0
Operational
GCP US Central - prod-us-central-0
Operational
GCP US Central - prod-us-central-3
Operational
GCP US Central - prod-us-central-4
Operational
GCP US East - prod-us-east-1
Operational
play.grafana.org
Operational
Federal Cloud - AWS US Gov West
Operational
Grafana Cloud: Graphite
Operational
AWS Australia - prod-ap-southeast-2: Querying
Operational
AWS Australia - prod-ap-southeast-2: Ingestion
Operational
AWS Brazil - prod-sa-east-1: Querying
Operational
AWS Brazil - prod-sa-east-1: Ingestion
Operational
Resolved -
This incident has been resolved.
May 1, 10:27 UTC
Monitoring -
A fix has been implemented and we are monitoring the results.
May 1, 09:43 UTC
Update -
We are continuing to investigate this issue.
May 1, 09:42 UTC
Investigating -
We are investigating an issue with ingesting Metrics for AWS Cloud Provider Observability with Metric Streams. Users experiencing this issue may encounter ingestion errors in the "prod-us-central-7" region only starting from ~06:30UTC.
Engineering is actively engaged and assessing the issue. We will provide updates accordingly.
May 1, 09:14 UTC
Resolved -
After further review, this was a false alarm and should not have affected any users.
This incident has been resolved. Thank you for your patience.
Apr 30, 15:11 UTC
Investigating -
Successful requests have dropped, users may not be able to access their instances.. The issue is under investigation.
Apr 28, 09:20 UTC
Resolved -
This incident has been resolved. Thank you for your patience.
Apr 29, 13:37 UTC
Monitoring -
We’ve identified the cause of the issue impacting SQL datasources. Our team is currently implementing a fix and are monitoring the results to confirm the issue is fully resolved. Services may start to recover during this time.
Apr 28, 18:59 UTC
Investigating -
We are currently investigating reports of issues affecting SQL-based data sources where users are unable to save credentials.
This appears to impact a subset of customers and may be occurring across multiple regions. We are actively working to determine the scope and root cause.
We will provide updates as more information becomes available.
Apr 28, 18:46 UTC
Resolved -
We experienced degraded performance affecting Performance Testing from 13:10 UTC to 13:20 UTC. During this time, users may not have been able to start new test runs.
The issue has been resolved, and the service is now operating normally.
We apologize for any disruption this may have caused and appreciate your patience.
Apr 29, 12:00 UTC
Resolved -
We were facing an incident with AWS Metrics Streaming integration in us-east-3 region manifesting in elevated ingestion latency. The incident started at around 10:45 UTC and was resolved at around 12:30 UTC. Some tenants could see an elevated write latency, but all requests were being processed and we don't expect any data loss during the time of the incident.
The incident is now resolved, but we keep monitoring the system's health.
Apr 29, 10:30 UTC
Resolved -
This incident has been resolved. Thank you for your patience.
Apr 27, 23:24 UTC
Monitoring -
We’ve implemented a fix and are monitoring the results to confirm the issue is fully resolved. Services may start to recover during this time.
Apr 27, 23:13 UTC
Identified -
We’ve identified the cause of the issue impacting the InfluxDB datasource. Our team is currently implementing a fix.
Apr 27, 18:01 UTC
Investigating -
We’re currently investigating an issue affecting the InfluxDB plugin. Some users may see intermittent failures.
Our team is actively working to identify the cause. Thank you for your patience.
Apr 27, 17:08 UTC
Resolved -
Grafana Labs has taken steps to safeguard the Grafana Cloud platform against the distribution of unauthorized emails. We have implemented the following changes to new Grafana Cloud Free and Trial accounts, effective immediately:
Users who open Grafana Cloud Free and Trial accounts can only send email alerts and reports to users within their Grafana instance. Other email recipients will be rejected. Additionally, the Cloud Alertmanager is no longer available for these instances, requiring all alerting to be configured via the native Grafana Alertmanager. All other integrations for alerting remain functional.
If users would like to expand their email capabilities, they can upgrade their Grafana Cloud account to scale their use case, as needed.
These changes to alerting and reporting will be applied to all new Grafana Cloud Free and Trial accounts. Existing Grafana Cloud Free and Trial accounts opened before April 20 and all other Grafana Cloud account types are unaffected by these changes.
Apr 24, 15:04 UTC
Update -
Grafana Labs is implementing measures to safeguard the Grafana Cloud platform against ongoing unauthorized use while preserving the capabilities relied upon by our community. Effective immediately, we have made the following modifications to the platform:
Alerting Email alerting has been disabled for new Grafana Cloud Free and Trial accounts; however, all other integrations such as webhooks remain functional. Additionally, Cloud Alertmanager is now disabled for Grafana instances in these accounts, requiring all configuration to occur via the native Grafana Alertmanager. Existing Grafana Cloud accounts remain unaffected by these restrictions and all other Grafana Cloud account types are also unaffected.
Reporting To prevent the distribution of unauthorized emails, Grafana instances in new Grafana Cloud Free and Trial accounts are limited to sending reports exclusively to users within their Grafana instance. Standard reporting functionality continues unrestricted for existing Grafana Cloud accounts and all other Grafana Cloud account types.
We remain committed to further refining platform security to ensure a safe and open environment for our entire user base.
Apr 22, 15:03 UTC
Update -
We are continuing to monitor for any further issues.
Apr 20, 22:07 UTC
Monitoring -
Grafana Labs is taking steps to safeguard our Grafana Cloud platform against unauthorized use while maintaining the Grafana Cloud Free and Trial tiers of service our users and the community have come to rely on. As of Monday April 20, alerting and reporting capabilities have been disabled in new Grafana Cloud Free and trial stacks. We are working towards deploying improvements and restoring those functionalities in a way that keeps our platform secure and open for all of our users.
Apr 20, 21:12 UTC
Resolved -
This incident has been resolved. Thank you for your patience.
Apr 23, 20:07 UTC
Update -
We are continuing to monitor for any further issues.
Apr 20, 15:08 UTC
Update -
We have deployed mitigation and seen improvement in write failures over the past week. We are still seeing intermittent spikes in latency and continue to monitor.
Apr 14, 20:11 UTC
Update -
We are still seeing intermittent issues and continue to seek a resolution
Apr 8, 20:32 UTC
Update -
We are continuing to monitor for any further issues.
Apr 2, 21:38 UTC
Update -
We are continuing to monitor this through the weekend.
Mar 27, 21:05 UTC
Update -
We are continuing to monitor the previously impacted environments.
Mar 26, 17:45 UTC
Monitoring -
A fix has been implemented and we are monitoring the results.
Mar 26, 12:04 UTC
Update -
We are continuing to investigate this issue.
Mar 25, 21:35 UTC
Resolved -
This incident has been resolved. Thank you for your patience.
Apr 23, 20:01 UTC
Monitoring -
We’ve implemented a fix and are monitoring the results to confirm the issue is fully resolved. Services may start to recover during this time.
Apr 23, 14:39 UTC
Investigating -
We’re currently investigating an issue affecting Cloudwatch datasources. Our team is actively working to identify the cause. Thank you for your patience.
Apr 23, 14:26 UTC
Resolved -
This incident has been resolved. Thank you for your patience.
Apr 20, 14:30 UTC
Update -
The issue is now confirmed to be widespread, affecting Prometheus across all regions.
Customers may continue to experience elevated 429 (rate limit) errors, particularly when querying metrics, with failures or inconsistent responses possible.
Our engineering team remains fully engaged and is actively working on mitigation and resolution efforts with the highest priority.
Apr 20, 14:21 UTC
Investigating -
We are currently experiencing a major incident causing elevated 429 (rate limit) errors across multiple regions, primarily impacting metrics querying.
This is a high-priority issue, and our engineering team is actively engaged and working urgently to identify the root cause and restore full service as quickly as possible. Customers may experience widespread failures or delays when querying metrics during this time.
We understand the significant impact this may have and will continue to provide updates as more information becomes available.
Apr 20, 14:09 UTC
Resolved -
This incident has been resolved
Apr 17, 22:58 UTC
Monitoring -
Currently prod-us-east-0 and prod-eu-west-3 have recovered, and we are continuing to monitor prod-us-central-0 which is in the process of recovery.
Apr 17, 22:09 UTC
Investigating -
As of 20:52 UTC, we are currently investigating degraded Query Caching performance in multiple regions. For datasources where query caching is configured, some queries may take longer than usual.
Our team is actively working to identify the cause. Thank you for your patience.
Apr 17, 21:23 UTC