If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below. |
|
|
|
Thread Tools | Display Modes |
|
#1
|
|||
|
|||
Long term archival storage
I have around 20 Tb of data that I (a) want to store for a very (50
years)long time and also have available for search and download. The data consists of two types: (a) the preservation masters which is the data we want to keep and is in tiff and bwf formats among others (b) the viewing copies which are in derived formals such as png and mp3. I am coming down to an HSM type of solution which a large enough front end cache to allow us to keep the viewing copies online at all times but which allows the archival copies to disappear off to tape to be cloned and duplicated etc. Anyone else doing this? Anyone got a better idea? -dgm |
#2
|
|||
|
|||
|
#3
|
|||
|
|||
depends, a lot of companies have regulatory requirements that data be
kept for a long time. If it needs to be accessible quickly, the ATA drive solution is advisable. I think the cost is around $7 per GB for stuff like netapp nearstore solutions. The particular product which would simulate a worm drive is called snaplock. |
#4
|
|||
|
|||
Paul Rubin wrote: [...] Sounds kind of complicated. Where's this data now, how is it stored, and how fast are you adding to it and through what kind of system? 20 TB isn't really big storage these days. You could have a small tape library online and move incoming raw data to tape immediately while also making the online viewing copies on disk. HSM systems with automatic migration and retrieval are probably overkill. It is kind of complicated. Currently we have 6Tb digitised and are adding 0.1Tb/week. Now this is data that's stuff that needs to be kept for ever - the audio stuff is world heritage stuff. The driver for using HSM is two fold 1) keeping multiple copies securely including offsite 2) we know we have a 900kg gorilla called video waiting in the wings .... |
#5
|
|||
|
|||
50 years is a long time by archival standards. Most media/drive
manufacturers guarantee 30 years under nominal conditions. You'll have to store your masters under optimal conditions and check them regularly (read them every couple of years, check error rate, copy if needed etc.) to go beyond that. Which raises the question where you'll find the equipment (spares, service) a couple decades down the road. Example from my perspective: If you have DLTtape III tapes lying around from the early nineties, you'd better do something about them now since Quantum EOL'd the DLT8000, which is the last generation that will read those tapes. Service will be available for another 5 years. That's 20 of the 50 years. Basically you'll have to copy the data to state of the art media about every decade. Don't try to store spare drives for the future, that doesn't usually work - electromechanical devices age when they're not in use too. There have been numerous stories about the problems NASA has retrieving old data recordings. Your project will face the same. Fortunately 20 TB isn't a big deal any more and will be less so in the future. The front end doesn't really matter, but the archive will need a lot of thought and care. Think what the state of the art was 50 years ago. |
#6
|
|||
|
|||
"RPR" wrote in message
oups.com... Basically you'll have to copy the data to state of the art media about every decade. Don't try to store spare drives for the future, that doesn't usually work - electromechanical devices age when they're not in use too. In addition to Ralf-Peter's comment, you better think long and hard about how you will be accessing that data 50 years from now, from an application point of view. 50 years from now, the computing devices will be radically different from today's PC's. Unless you have documented every bit about the format of the files you stored and the environment you need to recreate the information, even migration to state of the art media will not help. Consider a Word Perfect 4.2 file from 20 years ago. You'll need some effort today to open and read such a file. Because the format is relatively simple, you can still read the text using any hex editor. But recreating the page formatting maybe harder already. Now consider your MP3 and picture files which are heavily encoded en compressed, and fast forward to the year 2055. Unless you know exactly how they are recreated, all you'll have 50 years from now is a bunch of zeroes and ones. This is scary for single files, but things are even worse when multipple files form a single context. Think databases with external pointers. Think HTML files with web links. How much of that will exist 50 years from now? For permanent long-term records, store the information on a medium that can be interpreted by the most universal and long-term computer you have - the one between your ears -. Microfiche and dead trees aren't obsolete just yet... Rob |
#7
|
|||
|
|||
On Tue, 29 Mar 2005 23:36:05 +0200, "Rob Turk"
wrote: "RPR" wrote in message roups.com... Basically you'll have to copy the data to state of the art media about every decade. Don't try to store spare drives for the future, that doesn't usually work - electromechanical devices age when they're not in use too. In addition to Ralf-Peter's comment, you better think long and hard about how you will be accessing that data 50 years from now, from an application point of view. 50 years from now, the computing devices will be radically different from today's PC's. Unless you have documented every bit about the format of the files you stored and the environment you need to recreate the information, even migration to state of the art media will not help. Consider a Word Perfect 4.2 file from 20 years ago. You'll need some effort today to open and read such a file. Because the format is relatively simple, you can still read the text using any hex editor. But recreating the page formatting maybe harder already. Ok so a lot of converters do an incomplete job, but is this really so complicated? Save a copy of the application(s) and maybe the OS that ran it with the data. Between backwards compatibility and improving emulation technology it might be more doable than you think. Also keeping data for 50 years doesn't necessarily imply keeping storage devices for 50 years. Periodic upgrades of the storage and maybe even the file format of the data might be what needs to happen to realistically keep useable information for many decades. A major overhaul like this around every 10 years seems to be working for me pretty well. Waiting 15 years or more tends to be problematic. Your mileage may vary and, well, the past is not always a good indicator of the future. |
#8
|
|||
|
|||
Curious George writes:
Consider a Word Perfect 4.2 file from 20 years ago. You'll need some effort today to open and read such a file. Because the format is relatively simple, you can still read the text using any hex editor. But recreating the page formatting maybe harder already. Ok so a lot of converters do an incomplete job, but is this really so complicated? Save a copy of the application(s) and maybe the OS that ran it with the data. Between backwards compatibility and improving emulation technology it might be more doable than you think. I would say that most of these conversion problems have stemmed from secret, undocumented formats. Formats like jpg and mp3, which are well documented and have reference implementations available as free source code, should be pretty well immune to the problems. |
#9
|
|||
|
|||
The point about file formats is well made, but we've been through the
same arguement in detail already. We're choosing file formats which are publically described for which there are multiple (open source) clients. The idea is to be able to ensure that we have the format description and enough example code to be able to recreate viewers in the future. That's why we're using tiff and bwf as the archival masters. I don't care about the mp3's as they are *derived* copies - we can as easily use ogg vorbis, or whatever we're using in 2055 as long as we can parse the original compression free datastream |
#10
|
|||
|
|||
|
|
Thread Tools | |
Display Modes | |
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Long term 100% CPU | 'Captain' Kirk DeHaan | Dell Computers | 6 | December 15th 04 06:45 AM |
Redundant network storage | Andrew Gideon | Storage & Hardrives | 0 | December 6th 04 05:12 PM |
SAN (Storage Area Network) Security FAQ Revision 2004/06/23 - Part 1/1 | Will Spencer | Storage & Hardrives | 0 | June 23rd 04 07:04 AM |
Enterprise Storage Management (ESM) FAQ Revision 2004/06/23 - Part 1/1 | Will Spencer | Storage & Hardrives | 0 | June 23rd 04 06:58 AM |
Get the Serial Number with Visual Basic | Michael Wittmann | General | 15 | November 15th 03 06:03 PM |