Re: Backup - Mailing list pgsql-general

From Achilleas Mantzios
Subject Re: Backup
Date
Msg-id f1a0c76b-c645-4530-b768-fea480eab56e@cloud.gatewaynet.com
Whole thread Raw
In response to Re: Backup  (Ron Johnson <ronljohnsonjr@gmail.com>)
Responses Re: Backup
List pgsql-general


Στις 16/10/24 22:55, ο/η Ron Johnson έγραψε:
On Wed, Oct 16, 2024 at 3:37 PM Andy Hartman <hartman60home@gmail.com> wrote:
I am very new to Postgres and have always worked in the mssql world. I'm looking for suggestions on DB backups. I currently have a DB used to store Historical information that has images it's currently around 100gig. 

I'm looking to take a monthly backup as I archive a month of data at a time. I am looking for it to be compressed and have a machine that has multiple cpu's and ample memory.

Suggestions on things I can try ?  
I did a pg_dump using these parms
--format=t --blobs lobarch

it ran my device out of storage:

pg_dump: error: could not write to output file: No space left on device

I have 150gig free on my backup drive... can obviously add more

looking for the quickest and smallest backup file output...

Thanks again for help\suggestions 

Step 1: redesign your DB to NOT use large objects.  It's an old, slow and unmaintained data type.  The data type is what you should use.
You mean bytea I guess. As a side note, (not a fan of LOs), I had the impression that certain drivers such as the JDBC support streaming for LOs but not for bytea? It's been a while I haven't hit the docs tho.

Step 2: show us the "before" df output, the whole pg_dump command, and the "after" df output when it fails. "du -c --max-depth=0 $PGDATA/base" also very useful.

And tell us what version you're using.

--
Death to <Redacted>, and butter sauce.
Don't boil me, I'm still alive.
<Redacted> crustacean!

pgsql-general by date:

Previous
From: Ron Johnson
Date:
Subject: Re: Backup
Next
From: Ron Johnson
Date:
Subject: Re: Backup