Hello Ben!
On Dienstag, 3. Juli 2007, Ben Collins-Sussman wrote:
> On 7/3/07, Ph. Marek <philipp.marek@bmlv.gv.at> wrote:
> > I'd like to push the contents of some repository (say, 2GB) to many
> > (order of several hundred) machines at once, with a granularity of, say,
> > a day.
> >
> > Believing (but not having benchmarked) that restoring file data from a
> > compressed repository is *much* more CPU-using than simply streaming from
> > a RAID array, I thought that the easy way would be to generate a tar
> > file, and let the clients fetch it.
>
> If the repository itself is accessible, you can just 'svnadmin dump'
> and let clients fetch that. A dumpfile is extremely similar to a
> tarfile.
>
> If you're already planning on making several hundred machines "pull"
> repository data, why not have each of them just use svnsync to
> replicate (pull) the repository to their local disks? What's great
> about that is that only the most recent commits will be replicated
> each day. Clients can then run 'svn export' from their local
> repositories.
The clients need just the current snapshot, not the history.
And this is for the first setup, so the clients have no "old" version.
> Or, if you want to avoid the CPU work of actually interpreting
> (decompressing) the repository data,
Yes, that's what I want!
> just have each client machine use
> rsync to do daily pulls instead.
That's not what I want, because
-) they have no old version, so
-) rsync would have to push *all* files, anyway, and
-) as mentioned, a nice big tar file can be streamed very easily ...
With a directory structure of some hundred thousand files you can
get lost in seeks.
Thank you, anyway.
Regards,
Phil
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscribe@subversion.tigris.org
For additional commands, e-mail: dev-help@subversion.tigris.org
Received on Tue Jul 3 13:22:21 2007