rsync alternative? (too many files)

Tony Godshall togo at of.net
Mon Mar 7 15:11:03 PST 2011


On Sun, Mar 6, 2011 at 15:19, Seth David Schoen <schoen at loyalty.org> wrote:
> Tony Godshall writes:
>
>> > find . -type f -print0  -xdev | xargs -0
>> > ...running cp piped into ssh, or whatever.  'Ware slowness.
>>
>> Yeah.  I looked at doing a find ... -type d -print0 | xargs -0 mkdir
>> followed by a one that does the rsyncs without the recursion[1] so
>> that each rsync would have only one file to do, but that doesn't,
>> unless I'm missing something, preserve the hardlinks, which is pretty
>> important since I've got something like 2.5TB residing in about 1.5TB
>> after file-level deduplication that I'm trying to copy to a 2TB
>> removable volume.

[Seth]
> If you're sure that the filenames don't contain tabs, you can...

Hi Seth.

I must not have expressed myself clearly.

There are excessive unique files, not duplicate entries in a list of files.

The files have already been deduplicated in the sense that entries to
files containing the same content are hardlinks.

If I were to copy the files to new media without retaining the
hardlinks, they would take up way more space.

Which is why I was unable to take Rick's otherwise very good suggestion.

That is what I was trying to express.  Am I being clearer now?

I may have to go hack on rsync... :-(

[Don]
> Most of the GNU utilities have an option that will
> use a null byte instead of a linefeed as a delimiter. ...

Yes, Rick had that in his solution.  A very useful feature that I use
all the time.

Tony


More information about the bad mailing list