Oracle FAQ | Your Portal to the Oracle Knowledge Grid |
Home -> Community -> Usenet -> c.d.o.server -> Re: export table > 2G
The other thing we used to do, particularly with version 6 was to use dd to
write appropriate numbers of blocks and keep generating a new file each time
dd had produced a file. Thus, the one compressed input stream (from a named
pipe from export) could be broken up into multiple files. Import was easy,
'cause they simply got cat'ed together, then to uncompress and then to the
named pipe.
"Gollum" <gollum_nospam_at_worldonline.dk> wrote in message
news:7tXg6.27456$fa3.1675814_at_news010.worldonline.dk...
> In fact it is even easier to:
>
> $ exp ... file=stdout ... | gzip > expfile.dmp.gz
>
> Unfortunately this does not work for import (ie. using file=stdin), so in
> that case you will need the named pipe.
>
> HTH,
> Gollum
>
> <jdarrah_co_at_my-deja.com> wrote in message
> news:96190g$8du$1_at_nnrp1.deja.com...
> > If your using 8i you can export to multiple files. otherwise on unix
> > export the file to a pipe then gzip -c from the pipe to a file
> >
> > $ mknod exppipe.dmp p
> > $ exp file=expipe.dmp
> > from another shell
> > $ gzip -c expipe.dmp > expfile.dmp.gz
> >
> > This should work as long as the table is under around 14G
> >
> > In article <961854$7d1$1_at_nnrp1.deja.com>,
> > rkatofiasc_at_classifiedventures.com wrote:
> > > I need to export a table that is over 2G. How can I get around
oracle's
> > > limitation?
> > >
> > > Thanks Rob
> > >
> > > Sent via Deja.com
> > > http://www.deja.com/
> > >
> >
> >
> > Sent via Deja.com
> > http://www.deja.com/
>
>
Received on Fri Feb 09 2001 - 21:07:38 CST