Render Outage History
Past incidents and downtime events
Complete history of Render outages, incidents, and service disruptions. Showing 50 most recent incidents.
February 2026(3 incidents)
Degraded deploys in Ohio
1 update
Deploys are experiencing degraded performance and may take longer to complete.
Degraded Deploys in Singapore Region
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are currently investigating this issue.
Degraded deploys in all regions
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are currently investigating this issue.
January 2026(5 incidents)
External connectivity issues with Postgres databases hosted in Singapore
4 updates
This incident has been resolved. Please reach out to support@render.com for any follow-up questions.
A fix has been implemented and we are monitoring the results.
The issue has been identified and a fix is being implemented.
We are currently investigating this issue.
Metrics impacted for some services in Oregon
4 updates
This incident has been resolved.
Metrics for impacted services in Oregon are now being displayed. Metrics will be missing from impacted services from 2026-01-23 00:50 to 2026-01-23 01:10 UTC
The issue has been identified and a remediation is being implemented
Metrics for some services in Oregon are currently impacted and may not be displaying.
Delays in starting instances on services
2 updates
Instance creation times have been restored to expected timerames. This issue has been resolved.
High demand for new instances has created a backlog for some services in the Oregon region. Services attempting to add new instances including those for new deploys, instance scale ups, restarts, etc. may see delays doing so.
Some application and build logs are missing on the dashboard
5 updates
This incident has been resolved.
We’re seeing steady recovery now, and logs should be showing again. We are still monitoring to confirm the longer-term recovery.
We’ve identified the issue and are now in recovery. Recovery may be slow due to the large volume of logs involved.
We believe we’ve identified the root cause of the issue, and we’re currently doing some additional investigation to make sure it’s resolved properly.
Some logs, especially build and application logs, may be temporarily missing. We’re actively investigating this and will work to fix it as soon as possible. Builds can still complete successfully even if the logs aren’t showing up.
Deploy delays in Oregon
3 updates
This incident has been resolved.
We have implemented a fix and are monitoring for further issues.
Some users may experience slower build times for services deployed in Oregon.
December 2025(6 incidents)
Deploy delays in Virginia
3 updates
Deploy performance has returned to expected levels.
A fix has been implemented and we are monitoring the results.
We are currently investigating this issue.
Unable to view service events
2 updates
This incident has been resolved.
When viewing service events, an error is returned. We are currently investigating the issue.
Elevated Latency for Requests to Web Services and Static Sites in Frankfurt
4 updates
This incident has been resolved.
Latency has remained stable. We continue to monitor the situation.
Latency has normalized. We continue to investigate with our upstream vendor to identify the cause.
We are currently investigating this issue.
Services not accessible
5 updates
From 08:47 to 09:11 UTC, all incoming web traffic in every region failed to reach services and returned 500 errors instead. Our dashboard and API were down too. Background workers, private services, and cron jobs were not affected. The upstream provider has recovered now, and we’re no longer seeing any issues on our side.
The upstream provider is recovering, and we’re seeing recovery on our side too.
Access to services is now recovering, and we are continuing to monitor.
We're experiencing issues with an upstream provider.
We're investigating services not being accessible
Increased Latency in Updates to Oregon Services
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
Creation of services or changes to existing services hosted in our Oregon region are experiencing increased latency. We are currently investigating.
Custom Domains: New certificates stuck on pending
3 updates
We understand the issue is resolved now. If you're still seeing issues, please reach out.
The provider is actively working on the issue and we’re seeing some progress on certificate issuance. We’re still waiting on full confirmation that the fix is complete.
You may see certificates stuck on 'Pending' after adding a custom domain. We’ve located an issue with a provider and are looking into it right now.
November 2025(8 incidents)
Web services (Oregon) and static sites availability disruption
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results. Impact resulted in intermittent latency, timeouts, and errors for some services for ~6 minutes (11:46-11:52 PST).
We are currently investigating an issue impacting web service and static site availability.
Increased slowness in Dashboard
4 updates
The incident has been resolved.
Dashboard performance remains healthy and we continue to monitor.
Dashboard performance has recovered. We are continuing to investigate the root cause.
We are currently investigating this issue.
Elevated rates of deploy failures
4 updates
This incident has been resolved.
The issue has been identified and a fix is being implemented.
We are continuing to investigate this issue.
We are currently investigating this issue.
GitHub-backed services failing to build in all regions
4 updates
This incident has been resolved.
The upstream provider has implemented a fix and recovery is ongoing. We are continuing to monitor the situation.
An upstream provider is experiencing an outage. We are monitoring the situation.
We are currently investigating this issue.
An upstream provider major incident is affecting some Render services
4 updates
We have observed no further impact and the upstream provider has affirmed full resolution.
The upstream provider has resolved the issue. We’re still checking to see if there’s any remaining impact on our side.
The upstream provider is still suffering from the incident, and we are still waiting for further mitigations from them.
We’re aware of a major incident with an upstream provider that’s impacting some services on Render. You might see some 500s until it’s resolved upstream. We’re also investigating on our side.
Metrics/Logs missing for Oregon services
2 updates
This incident has been resolved.
Metrics and Logs for services hosted in Oregon are missing due to a platform incident. We are working to resolve this issue now.
Cron Job runs cannot be cancelled from our dashboard or the API
2 updates
This incident has been resolved.
We’re looking into why this is happening. The cancel button on a run doesn’t actually stop it right now. The current workaround is to suspend and then unsuspend the cron to force-cancel the run. If that doesn’t do the trick, please reach out to our support team.
Increased 404s in Oregon (Web Services) and Static Sites
9 updates
# Summary As an infrastructure provider, providing a reliable platform that allows our customers to build and scale their applications with confidence is our highest obligation. We invest heavily to ensure our platform is highly reliable and secure, including in our routing layer that handles billions of HTTP requests every day. On November 5, 2025, we inadvertently rolled back a performance improvement that was gated behind a feature flag. This led to disruption in the form of intermittent 404s for some web services and static sites deployed to the Oregon region. We have fully identified the sequence of events that led to this outage and are in the process of taking steps to prevent it from recurring. # Impact There were two periods where some customers hosting web services and static sites in the Oregon region experienced a partial outage with intermittent 404s. The first period occurred between 10:39 AM PST and 11:25 AM PST . At this time, two Render clusters had slightly degraded service. One cluster returned a negligible number of 404 responses, and the other cluster returned 404 responses for approximately 10% of requests. The second period occurred between 11:59 AM PST to 12:34 AM PST and saw more significant service degradation. During this period, about 50% of all requests to services in the affected cluster received a 404 response. All newly created services in these clusters were affected and received 404 responses during the incident. Updates to existing services were also slow to propagate. Free tier services that were recently deployed or waking from sleep were also affected. # Root Cause Render's routing service depends on a metadata service to receive information about the user services it routes traffic to. When the routing service first starts and upon occasional reconnection, it will request and receive a large volume of data from the metadata service. Earlier in 2025, we successfully deployed a memory optimization related to data transfer between the metadata and routing services using a feature flag. In late October, we removed the flag from code and redeployed, but we didn't redeploy the metadata service, which still depended on the flag. On November 5th, we cleaned up unreferenced feature flags from our system. This caused the metadata service to revert to its less efficient data transfer method, leading to memory exhaustion and crashes. Our routing service is designed to handle metadata service outages and continue serving traffic based on its last known state. However, newly created instances that could not load their initial state were incorrectly sent requests, resulting in 404 errors. During the first period of impact, the metadata service was crashing in two of our clusters, and only a small fraction of routing service instances were impacted. During the second period of impact, we saw a large increase in HTTP requests for services in the affected cluster. This triggered scale-ups of the routing service, all of which returned 404 errors. # Mitigations ## Completed * Increased memory available to the metadata service \(this has since been reverted\) * Temporarily re-enabled the feature flag to support more efficient data transfer between the routing and metadata services \(this has since been removed\) * Deployed the metadata service to no longer rely on the feature flag * Enhanced our monitoring of the metadata service to alert us of this particular failure mode ## Planned * Improve our feature flag hygiene practice to prevent the removal of a feature flag while it is still being evaluated * Prevent the routing service from receiving traffic if it never successfully loaded state from the metadata service
This incident has been resolved.
We are continuing to monitor for any further issues.
A fix has been implemented and we are monitoring the results.
We are continuing to work on a fix for this issue.
We have identified continuing issues in Oregon. A fix is being worked on.
A fix has been implemented and we are monitoring the results.
The issue has been identified and a fix is being implemented.
We are currently investigating the issue.
October 2025(8 incidents)
Failure to spin free web services back up after inactivity
1 update
Between 2025-10-28 at 17:00 UTC and 2025-10-30 at 17:22 UTC, a change was active that caused some free web services to fail to spin back up after inactivity. Most free web services were unaffected. While the change was reverted, any services that remain impacted should redeploy to resolve.
Degraded builds and deploys in Virginia
3 updates
This incident has been resolved.
An upstream provider is experiencing issues provisioning infrastructure. We continue to monitor the situation. Paid services are experiencing less delay than free services.
Builds and deploys may be slower than usual. We are currently investigating this issue.
Pre-deploys are failing in some regions
6 updates
This incident has been resolved.
A fix has been rolled out, and a re-deploy should now work as expected. We’re still keeping an eye on how the fix performs though.
We’ve found the root cause and are rolling out a fix.
Some pre-deploys in Singapore are also affected.
We think some pre-deploys are failing in Oregon and Frankfurt. We haven’t found any other affected regions so far, but we’re actively checking.
We are currently investigating this issue.
An upstream provider is experiencing some issues that are affecting parts of our platform (Virginia)
11 updates
This incident has been resolved.
All Render services have recovered, our upstream provider is continuing to recover. We are continuing to monitor impact.
Web Services (paid and free) and Static Site request latencies have returned to normal levels. Issues involving PostgreSQL database creation and inability to create backups persist.
The upstream provider has not yet recovered. We are still seeing request latency for Web Services and Static Sites in Virginia, and some users are unable to create new databases or backups.
Requests routed to Web Services have begun experiencing issues.
We’re seeing some issues again with a few components. Database creation might be slow in Virginia or appears stuck during the creation process.
We’re no longer seeing any issues related to Postgres databases from this incident on our platform.
We are continuing to monitor for any further issues.
We’re seeing steady recovery and keeping an eye on all components to make sure everything’s fully caught up. The upstream provider is still going through its own recovery process too.
Several of our tools were also affected during that time, including support tools, so responses may have been delayed or missed between 08:00 and 09:30 UTC. We’re working through the requests as quickly as we can.
We started seeing increased errors in our infrastructure around 08:00 UTC. Parts of our platform were affected by an outage with an upstream provider. We know that new database creation and backup creation were impacted, but we’re still assessing if there’s any broader impact. We’re seeing signs of recovery now, but we’re continuing to monitor.
Incorrect IP allowlists configured for new Environments created via REST API
2 updates
Changes were deployed to fix the issue with new Environments created via the REST API. All affected Environments have been updated to be their default Allow-All if not otherwise specified in the API call's parameters. This issue has been resolved.
We have identified and are working to fix Environments recently created via the REST API to ensure default IP allowlists are configured correctly. Until then, new Services created in these Environments may be responding to requests with unexpected errors.
Increased latency in Oregon region
3 updates
Latency has returned to baseline levels since 16:40 UTC and no further impact has been observed.
Peak impact occurred between 16:20 and 16:40 UTC. We are currently monitoring.
We are currently investigating increased latency in our Oregon region
Unable to create Postgres services or update their instance type in Oregon
3 updates
This incident has now been resolved. A subset of customers in Oregon, but not all, were impacted. Affected customers were unable to create Postgres services or update the instance type of Postgres services between 19:14 and 20:15 UTC.
A fix has been implemented and we are monitoring the results.
We are currently investigating this issue.
Partial degradation of service creation and deploys in Oregon
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are currently investigating this issue.
September 2025(9 incidents)
Small number of users impacted by stuck builds
4 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
The issue has been identified and a fix is being implemented.
We are aware of an issue resulting in stuck builds impacting a small minority of users with the "Wait" setting for their Overlapping Deploy Policy.
Image-based deploys failing due to upstream provider
3 updates
This incident has been resolved.
The upstream provider has rolled out a fix and are monitoring the issue. We are monitoring our systems as well.
Due to an outage from an upstream provider, users with image-based services are seeing failed deploys with reports of 401 errors.
Some Postgres databases can’t be created in Frankfurt
4 updates
This incident has been resolved.
We are continuing to work on the issue.
We’ve identified the issue, but we’re still investigating.
This doesn’t impact Postgres databases that are already running. It only partially affects Frankfurt. Any affected database that gets created will show a status of 'unknown'.
Dashboard operations degraded or failing
1 update
Dashboard operations were degraded for ~30 minutes, and within that period operations were mostly failing for ~5 mins.
Issues with deploys and spinning down free services in Virginia
1 update
Between 20:35 UTC and 20:53 UTC today, a process was unavailable that is responsible for handling port detection and deploys that were waiting due to a workspace's overlapping deploy policy. Queued deploys will proceed as expected in most cases. Additionally, free services may not have spun down on idle during that period.
Failure to pull some images from GitHub Container Registry
4 updates
We no longer see elevated failure rates for images pulled from GitHub Container Registry. If the problem persists please contact GitHub or Render's support team.
Intermittent failure continues to occur for images pulled from GitHub Container Registry. We recommend pulling the image locally and contacting GitHub if the issue persists locally. We are continuing to monitor the situation.
We have determined that this affects images from GitHub Container Registry only. We are continuing to investigate.
We are currently investigating this issue.
Deploy failure when using some public repos from GitHub
3 updates
This incident has been resolved.
Deploys are beginning to succeed when using a public repo and we are monitoring for any further issues.
Deploys may fail when using some public repos from GitHub. This affects some, but not all, public repos. We are investigating.
Some slack notifications failing August 1 - September 11
3 updates
We have completed monitoring our fix, and will publish an RCA attached to this incident.
Engineers have implemented a fix and are Monitoring.
We are investigating reports that a subset of Slack notifications were not being delivered. Fixing this has resulted in a high volume of notifications to be delivered. Some notifications may be delayed. We apologize for the noise that this may be creating in your configured channel(s).
Builds and Deploys erroring on Oregon services
3 updates
This incident has been resolved.
The root cause of these failures has been addressed and build failure rates have decreased substantially, we are monitoring for any further issues.
Services hosted in Oregon are seeing Builds and Deploys failing due to a platform issue. We have already begun steps to resolve this issue.
August 2025(9 incidents)
Incorrect bandwidth billing data in the Dashboard
3 updates
This incident has been resolved.
We have identified the issue, understand why it's occurring, and are working on a fix.
We are investigating reports of incorrect bandwidth billing data in the Dashboard.
Degraded network performance in Virginia region
4 updates
This incident has been resolved.
A fix has been implemented by our upstream provider and we are monitoring the results.
We have identified that this is tied to network performance issues from an upstream provider.
We are currently investigating this issue.
Degraded Webhooks and Queued Deploys
4 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are continuing to investigate this issue.
We are currently investigating this issue.
Delays in deploy completion
3 updates
This incident has been resolved.
Performance has returned to expected levels, we are monitoring for any further issues.
We have been alerted to performance issues causing delayed deployments for Oregon hosted services. Deploys are taking an increased amount of time to complete. We have begun steps toward remediating this issue.
Unable to issue certificate for wildcard custom domains
1 update
A change on our end inadvertently prevented wildcard custom domains from getting certificates. We've implemented and confirmed a fix. For any wildcard custom domains added to a service between approximately 2025-08-08T16:00Z and 2025-08-12T17:00Z, delete the custom domain and re-add it to the service.
Singapore region issues
3 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are currently investigating reports of issues with services in the Singapore region
Some services may have runtime errors
6 updates
This incident has been resolved.
We have not observed a recurrence of the issue after our fix and continue to monitor for any further issues.
A fix has been implemented and we are monitoring the results.
We are continuing to look into this and determine the best remediation strategy.
An underlying system package upgrade in our native environment runtime is causing a segfault for a small percentage of users. We are working on a fix.
We are currently investigating this issue.
Deploy failure on downloading bun
2 updates
This incident has been resolved.
Services, even those that don't use bun, may be experiencing deploy failure due to being unable to find bun to download. We're investigating the issue.
Point-In-Time Recovery restores degraded
1 update
Some Point-In-Time Recovery (PITR) backups of Render Postgres services hosted in Oregon began to error at approximately 10AM Pacific today (August 1). Attempts to restore backups from times later than 10AM may fail due to the inability to retrieve the data necessary to accomplish the restore. Restores for timeframes prior to 10AM today will succeed. We have already restored PITR coverage for the vast majority of Postgres services. Fewer than 10 services remain that may take until mid-day tomorrow for coverage to be restored.
July 2025(2 incidents)
The Render dashboard and REST API are slow
4 updates
This incident has been resolved.
A fix has been implemented and we are monitoring the results.
We are continuing to investigate this issue.
We are currently investigating this issue.
Inability to create or update Postgres services for some users in Oregon
2 updates
This incident has been resolved.
We have identified the issue, implemented a fix, and are monitoring for full restoration. This affects some users in the Oregon region, but not all.