Zero downtime migrations at Petabyte scale

planetscale.com

43 points by Ozzie_osman 3 days ago


Thaxll - an hour ago

We need more details on 6. This is the hard part, like you swap connection from A to B, but if B is not synced properly and you write to it then you start having diff between the two and there is no way back.

Like B is slightly out of date ( replication wise ) the service modify something, then A comes with change that modify the same data that you just wrote.

How do you ensure that B is up to date without stopping write to A ( no downtime ).

mattlord - 3 hours ago

Blog post author here. I'm happy to answer any related questions you may have.

WaitWaitWha - 2 hours ago

I split step 4 in their "high level, this is the general flow for data migrations".

4.0 Freeze old system

4.1 Cut over application traffic to the new system.

4.2 merge any diff that happened between snapshot 1. and cutover 4.1

4.3 go live

to me, the above reduces the pressure on downtime because the merge is significantly smaller between freeze and go live, than trying to go live with entire environment. If timed well, the diff could be minuscule.

What they are describing is basically, live mirror the resource. Okay, that is fancy nice. Love to be able to do that. Some of us have a mildly chewed bubble gum, a foot of duct tape, and a shoestring.

ksec - 2 hours ago

Missing 2024 in the Title.

redwood - 3 hours ago

Worth underlining that this is data migrations from one database server or system to another rather than schema migrations

clarabennett26 - 3 hours ago

[dead]