If you think having a backup is too expensive, try not having one

  • Thread starter Thread starter nsaspook
  • Start date Start date
AI Thread Summary
The South Korea data center fire has resulted in the potential loss of 858TB of government data due to a lack of backups, highlighting the critical importance of data protection strategies. A senior officer overseeing recovery efforts tragically died, underscoring the human impact of such data loss. The Ministry of Personnel Management is particularly affected, as it relied on a G-Drive system that failed to preserve eight years of work materials. Discussions emphasize the need for robust backup and restore strategies, as past experiences reveal that automated backups do not guarantee data recoverability. This incident serves as a stark reminder of the vulnerabilities in data management practices and the necessity for regular testing of recovery systems.
nsaspook
Science Advisor
Messages
1,486
Reaction score
4,931
https://www.datacenterdynamics.com/...-for-good-after-south-korea-data-center-fire/

858TB of government data may be lost for good after South Korea data center fire
Destroyed drive wasn't backed up, officials say
Meanwhile, a government worker overseeing efforts to restore the data center has died after jumping from a building.
As reported by local media outlet The Dong-A Ilbo, the 56-year-old man was found in cardiac arrest near the central building at the government complex in Sejong City at 10.50am on Friday, October 3. He was taken to hospital, and died shortly afterwards.
The man was a senior officer in the Digital Government Innovation Office and had been overseeing work on the data center network. His mobile phone was found in the smoking area on the 15th floor of the government building.

https://www.chosun.com/english/national-en/2025/10/02/FPWGFSXMLNCFPIEGWKZF3BOQ3M/

The Ministry of Personnel Management, where all affiliated officials use the G-Drive, is particularly affected. A source from the Ministry of Personnel Management said, “It’s daunting as eight years’ worth of work materials have completely disappeared.”
...
It provided 30GB, gigabytes, per public official. At the time, the Ministry of the Interior and Safety also issued guidelines to each ministry stating, “All work materials should not be stored on office PCs but should be stored on the G-Drive.”

G drive as in GONE drive.
1759972971676.webp
 
  • Sad
  • Like
  • Wow
Likes Greg Bernhardt, Rive and berkeman
Computer science news on Phys.org
Long ago, before backup utilities were so widely available, I worked in a computer lab that did backups to tape every night. The disk capacities were not big then. The administrators used a program that they wrote, which looped to copy every file to tape and finally rewind the tape for them to dismount. Every night, they started many backups for many computers and later dismounted all the tapes. They did that for a long time.
One day, an administrator noticed the tapes seemed to rewind many times in a single run. They saw that the program had the rewind inside the loop after every file copy. So only the last file remained on the backup tape and all others had been overwritten every night.

The moral of the story is to check that archived files can actually be retrieved.
 
Last edited:
  • Like
  • Haha
Likes russ_watters, jack action, PeroK and 1 other person
Years ago my wife and I arrived at Heathrow to find our airline was processing check ins on paper after a massive IT failure. The story we heard was it was due to a data center fire that had destroyed the primary system, and although they had had a backup it had been in the next room.
 
  • Wow
Likes FactChecker
In the latter part of my career, it became increasingly obvious that many data centre operations were dependent on the reliability of modern disk technology. They had a "backup strategy", of course, but little in the way of a "restore strategy". The first example I remember was in about 2005, where the tape storage facility was next to the disk storage units. And, there's was no way to manually extract tapes from this facility. A fire in that case could have destroyed everything, including all backups.

Restoring from tape became increasingly rare, with clever disk-based solutions like snapshots meaning that tape had become a last line of defence. Increasingly, the backups were completely automated, to the point where no one seemed to think very much about testing the restoration of data. It was assumed, more or less, that if the backups were running, then the restoration of data from them would be possible.

Generally, there was always a difficulty in organising a full restore from backup. If the test went wrong, then in principle the data was gone! Often a full restore was tested before a system went live - but once the system was live, it could take quite a bit of ingenuity to test a restore properly. In the old days, when a system required a single Unix server with a tape backup every night, it was relatively easy to restore onto a similar server somewhere else and test the restored system. It might take one techie less than a day to do the whole off-site restoration, using tapes that were stored off-site, where nothing from the live site was needed. In the 1990's we did this sort of thing regularly - although we always kept our fingers crossed when a restore from tape was involved.

But, as systems became increasingly complex and interdependent - with an environment of perhaps several hundred virtual servers - creating a test installation was a project in itself. In one of the last projects I worked on, it took about six weeks to configure the server infrastructure. If an additional user-test environment was required, then it took about six weeks for the various Windows, Linux, Oracle and other techies to do their thing. Then, all the application software had to be loaded. (I think it cost about £1 million per environment!). By contrast, in the old days, creating a new test environment could take a couple of hours - copy the executables, copy an existing database, do some configuration, update the backup schedule(!) and that was it. And, it came at no additional cost - as long as the server had enough memory for another Oracle database. In one case, we had a single Unix server with nine test enviornments on it.

I worked on a project in about 2012 where we did a major upgrade by going live on what had been the "disaster recovery" (DR) site and swapping the roles of live and DR data centres. That all went well, but it was a controlled project over several months and not a DR test as such.

In general, there seemed to be a strong reluctance to do a DR test, even where the facilities were available and an annual test was part of the contract. People always seemed to have better things to do! Also, there was a new generation of system support techies who had a different outlook on things. I was telling anyone who would listen that we were essentially dependent on the reliability of the technology. By 2014, when I retired, I strongly believed that a fire at a data centre would have been a real disaster, with little chance of full systems and data recovery.
 
Last edited:
  • Like
  • Informative
Likes berkeman, Ibix and FactChecker
I've also set up systems from scratch where my immediate boss seemed to think that just RAIDing the database server was enough. My argument that the eggs were still in one basket in case of fire or other force majerure went unheeded until I went over his head. That didn't make me popular, ironically.
 
  • Like
  • Sad
Likes nsaspook and FactChecker
This week, I saw a documentary done by the French called Les sacrifiés de l'IA, which was presented by a Canadian show Enquête. If you understand French I recommend it. Very eye-opening. I found a similar documentary in English called The Human Cost of AI: Data workers in the Global South. There is also an interview with Milagros Miceli (appearing in both documentaries) on Youtube: I also found a powerpoint presentation by the economist Uma Rani (appearing in the French documentary), AI...

Similar threads

Replies
1
Views
10K
Replies
8
Views
5K
Back
Top