An update on the continued cloud disruptions Netregistry customers are facing, our internal technical teams have worked in collaboration with the vendor and we've identified the root cause of the issue.
Frustratingly, although the 'why' has been identified, we can't at this point determine the trigger point, which is the 'how'.
As we have confirmed the root cause we are altering the architecture of our cluster to accommodate the limitation. This work will progress throughout the next 72 hours, and offers a medium term solution to the platform stability while we evaluate a long term resolution.
Tomorrow we are testing a medium term solution, although at this stage our focus is on building and testing to the needed standard before committing to production.
Again we know this has been a frustrating experience and we appreciate the impact it's had on you and your business. It's been trying at this end as well and we are optimistic that the current re-engineering will give us breathing space to work to a long term solution.
The current update is that we have good news and bad news.
The good news is that we have taken remediation work to mitigate risk of the loadbalancers crashing. This particular line of work started at around lunchtime yesterday and completed around 4pm. Since undertaking this the loadbalancers have stabilized and have operated normally over the last 18 or so hours.
The bad news is that the vendor has not yet delivered the hotfix for us to apply which wil hopefully address the long term stability. We're currently continuing to work with them, which presently includes them having sent an engineer onsite with our network team. As we get any further updates on this piece of work, including when we apply any patch which may impact customers we'll proactively announce via normal channels.
We're optimistic that in the short term we'll see ongoing stability.