Hello all, I will experiment with -Fc (custom). The file is already growing very large.
I am running this: ./pg_dump -t RECORDER -Fc postgres | gzip > /tmp/dump Are there any other options for large tables to run faster and occupy less disk space? Below is memory info: [root@onxl5179 tmp]# cat /proc/meminfo MemTotal: 16333720 kB MemFree: 187736 kB Buffers: 79696 kB Cached: 11176616 kB SwapCached: 2024 kB Active: 11028784 kB Inactive: 4561616 kB Active(anon): 3839656 kB Inactive(anon): 642416 kB Active(file): 7189128 kB Inactive(file): 3919200 kB Unevictable: 0 kB Mlocked: 0 kB SwapTotal: 33456120 kB SwapFree: 33428960 kB Dirty: 33892 kB Writeback: 0 kB AnonPages: 4332408 kB Mapped: 201388 kB Shmem: 147980 kB Slab: 365380 kB SReclaimable: 296732 kB SUnreclaim: 68648 kB KernelStack: 5888 kB PageTables: 37720 kB NFS_Unstable: 0 kB Bounce: 0 kB WritebackTmp: 0 kB CommitLimit: 41622980 kB Committed_AS: 7148392 kB VmallocTotal: 34359738367 kB VmallocUsed: 179848 kB VmallocChunk: 34359548476 kB HardwareCorrupted: 0 kB AnonHugePages: 3950592 kB HugePages_Total: 0 HugePages_Free: 0 HugePages_Rsvd: 0 HugePages_Surp: 0 Hugepagesize: 2048 kB DirectMap4k: 10240 kB DirectMap2M: 16766976 kB # CPUs=8 RHEL 6.5 The PG shared memory info is the defaults as I've not touched the .conf file. I am not a DBA, just a test tools developer who needs to backup the table efficiently. I am fairly new to PG and not an expert at Linux. Also if there are recommended backup scripts/cron that you recommend, please point them to me. Thanks!! On Thu, Oct 15, 2015 at 3:59 PM, Scott Mead <sco...@openscg.com> wrote: > > On Thu, Oct 15, 2015 at 3:55 PM, Guillaume Lelarge <guilla...@lelarge.info > > wrote: > >> 2015-10-15 20:40 GMT+02:00 anj patnaik <patn...@gmail.com>: >> >>> It's a Linux machine with 8 CPUs. I don't have the other details. >>> >>> I get archive member too large for tar format. >>> >>> Is there a recommended command/options when dealing with very large >>> tables, aka 150K rows and half of the rows have data being inserted with >>> 22MB? >>> >>> >> Don't use tar format? I never understood the interest on this one. You >> should better use the custom method. >> > > + 1 > > Use -F c > > > -- > Scott Mead > Sr. Architect > *OpenSCG* > PostgreSQL, Java & Linux Experts > > > http://openscg.com > > >> >> >>> -bash-4.1$ ./pg_dump -t RECORDER postgres --format=t -w > /tmp/dump >>> pg_dump: [archiver (db)] connection to database "postgres" failed: >>> fe_sendauth: no password supplied >>> -bash-4.1$ ./pg_dump -t RECORDER postgres --format=t > /tmp/dump >>> Password: >>> pg_dump: [tar archiver] archive member too large for tar format >>> -bash-4.1$ pg_dumpall | gzip > \tmp\db.out-`date +\%Y\%m\%d\%H`.gz >>> -bash: pg_dumpall: command not found >>> -bash: tmpdb.out-2015101510.gz: Permission denied >>> -bash-4.1$ ./pg_dumpall | gzip > \tmp\db.out-`date +\%Y\%m\%d\%H`.gz >>> >>> >>> Thank you so much for replying and accepting my post to this NG. >>> >>> On Thu, Oct 15, 2015 at 11:17 AM, Melvin Davidson <melvin6...@gmail.com> >>> wrote: >>> >>>> In addition to exactly what you mean by "a long time" to pg_dump 77k of >>>> your table, >>>> >>>> What is your O/S and how much memory is on your system? >>>> How many CPU's are in your system? >>>> Also, what is your hard disk configuration? >>>> What other applications are running simultaneously with pg_dump? >>>> What is the value of shared_memory & maintenance_work_mem in >>>> postgresql.conf? >>>> >>>> On Thu, Oct 15, 2015 at 11:04 AM, Adrian Klaver < >>>> adrian.kla...@aklaver.com> wrote: >>>> >>>>> On 10/14/2015 06:39 PM, anj patnaik wrote: >>>>> >>>>>> Hello, >>>>>> >>>>>> I recently downloaded postgres 9.4 and I have a client application >>>>>> that >>>>>> runs in Tcl that inserts to the db and fetches records. >>>>>> >>>>>> For the majority of the time, the app will connect to the server to do >>>>>> insert/fetch. >>>>>> >>>>>> For occasional use, we want to remove the requirement to have a server >>>>>> db and just have the application retrieve data from a local file. >>>>>> >>>>>> I know I can use pg_dump to export the tables. The questions are: >>>>>> >>>>>> 1) is there an in-memory db instance or file based I can create that >>>>>> is >>>>>> loaded with the dump file? This way the app code doesn't have to >>>>>> change. >>>>>> >>>>> >>>>> No. >>>>> >>>>> >>>>>> 2) does pg support embedded db? >>>>>> >>>>> >>>>> No. >>>>> >>>>> 3) Or is my best option to convert the dump to sqlite and the import >>>>>> the >>>>>> sqlite and have the app read that embedded db. >>>>>> >>>>> >>>>> Sqlite tends to follow Postgres conventions, so you might be able to >>>>> use the pg_dump output directly if you use --inserts or --column-inserts: >>>>> >>>>> http://www.postgresql.org/docs/9.4/interactive/app-pgdump.html >>>>> >>>>> >>>>>> Finally, I am noticing pg_dump takes a lot of time to create a dump of >>>>>> my table. right now, the table has 77K rows. Are there any ways to >>>>>> create automated batch files to create dumps overnight and do so >>>>>> quickly? >>>>>> >>>>> >>>>> Define long time. >>>>> >>>>> What is the pg_dump command you are using? >>>>> >>>>> Sure use a cron job. >>>>> >>>>> >>>>>> Thanks for your inputs! >>>>>> >>>>> >>>>> >>>>> -- >>>>> Adrian Klaver >>>>> adrian.kla...@aklaver.com >>>>> >>>>> >>>>> -- >>>>> Sent via pgsql-general mailing list (pgsql-general@postgresql.org) >>>>> To make changes to your subscription: >>>>> http://www.postgresql.org/mailpref/pgsql-general >>>>> >>>> >>>> >>>> >>>> -- >>>> *Melvin Davidson* >>>> I reserve the right to fantasize. Whether or not you >>>> wish to share my fantasy is entirely up to you. >>>> >>> >>> >> >> >> -- >> Guillaume. >> http://blog.guillaume.lelarge.info >> http://www.dalibo.com >> > >