On 05/18/2010 04:49 PM, aragonx@xxxxxxxxxx wrote: > Hello all, > > I need some ideas. > > I have a backup server that contains 10 ext3 file systems each with 12 > million files scattered randomly over 4000 directories. The files average > size is 1MB. Every day I expect to get 20 or so requests for files from > this archive. The files were not stored in any logical structure that I > can use to narrow down the search. This will be different moving forward > but it does not help me for the old data. Additionally, every day data is > added and old data is removed to make space. Do you know when data will be added or deleted? Is it randomly during the day or at a fixed time? > So, now that you know a little about the environment, I need ideas on how > to find the file I want to restore fast. > > Using find on the partition is slow. Have you thought about "locate" or any of its implementations (mlocate, slocate, etc)? It essentially indexes your disks for you and keeps the result in a database which it can search. Unlike "find", its not ATM, but, instead, when the disk was last indexed. So, it won't know about any changes made after the last index was done.... > I thought about using find and piping the output to a file. I started it > 50 minutes ago and it still isn't done on a single partition. Plus the > file is currently about 1.3GB and how would I maintain such a file? > > Would putting the file names + path in a database be faster? > > As always, any help would be greatly appreciated. > > --- > Will Y. -- Kevin J. Cummings kjchome@xxxxxxx cummings@xxxxxxxxxxxxxxxxxx cummings@xxxxxxxxxxxxxxxxxxxxxxx Registered Linux User #1232 (http://counter.li.org) -- users mailing list users@xxxxxxxxxxxxxxxxxxxxxxx To unsubscribe or change subscription options: https://admin.fedoraproject.org/mailman/listinfo/users Guidelines: http://fedoraproject.org/wiki/Mailing_list_guidelines