Commit fdf07fe1 authored by Peter Eisentraut's avatar Peter Eisentraut

For consistency with the rest of PostgreSQL, rename BLOBs to large objects

in messages and documentation.
parent 196700c3
<!-- <!--
$Header: /cvsroot/pgsql/doc/src/sgml/ref/pg_dump.sgml,v 1.36 2001/09/03 12:57:50 petere Exp $ $Header: /cvsroot/pgsql/doc/src/sgml/ref/pg_dump.sgml,v 1.37 2001/09/21 21:58:29 petere Exp $
Postgres documentation Postgres documentation
--> -->
...@@ -144,7 +144,7 @@ Postgres documentation ...@@ -144,7 +144,7 @@ Postgres documentation
<term>--blobs</term> <term>--blobs</term>
<listitem> <listitem>
<para> <para>
Dump data and <acronym>BLOB</acronym> data. Include large objects in dump.
</para> </para>
</listitem> </listitem>
</varlistentry> </varlistentry>
...@@ -616,8 +616,8 @@ connectDBStart() -- connect() failed: No such file or directory ...@@ -616,8 +616,8 @@ connectDBStart() -- connect() failed: No such file or directory
</para> </para>
<para> <para>
To dump a database called mydb that contains To dump a database called <literal>mydb</> that contains
<acronym>BLOB</acronym>s to a <filename>tar</filename> file: large objects to a <filename>tar</filename> file:
<screen> <screen>
<prompt>$</prompt> <userinput>pg_dump -Ft -b mydb &gt; db.tar</userinput> <prompt>$</prompt> <userinput>pg_dump -Ft -b mydb &gt; db.tar</userinput>
...@@ -625,8 +625,8 @@ connectDBStart() -- connect() failed: No such file or directory ...@@ -625,8 +625,8 @@ connectDBStart() -- connect() failed: No such file or directory
</para> </para>
<para> <para>
To reload this database (with <acronym>BLOB</acronym>s) to an To reload this database (with large objects) to an
existing database called newdb: existing database called <literal>newdb</>:
<screen> <screen>
<prompt>$</prompt> <userinput>pg_restore -d newdb db.tar</userinput> <prompt>$</prompt> <userinput>pg_restore -d newdb db.tar</userinput>
......
<!-- $Header: /cvsroot/pgsql/doc/src/sgml/ref/pg_restore.sgml,v 1.16 2001/09/13 15:55:24 petere Exp $ --> <!-- $Header: /cvsroot/pgsql/doc/src/sgml/ref/pg_restore.sgml,v 1.17 2001/09/21 21:58:30 petere Exp $ -->
<refentry id="APP-PGRESTORE"> <refentry id="APP-PGRESTORE">
<docinfo> <docinfo>
...@@ -213,7 +213,7 @@ ...@@ -213,7 +213,7 @@
<listitem> <listitem>
<para> <para>
Connect to database <replaceable class="parameter">dbname</replaceable> and restore Connect to database <replaceable class="parameter">dbname</replaceable> and restore
directly into the database. BLOBs can only be restored by using a direct database connection. directly into the database. Large objects can only be restored by using a direct database connection.
</para> </para>
</listitem> </listitem>
</varlistentry> </varlistentry>
...@@ -585,8 +585,8 @@ connectDBStart() -- connect() failed: No such file or directory ...@@ -585,8 +585,8 @@ connectDBStart() -- connect() failed: No such file or directory
<listitem> <listitem>
<para> <para>
<command>pg_restore</command> will not restore BLOBs for a single table. If <command>pg_restore</command> will not restore large objects for a single table. If
an archive contains BLOBs, then all BLOBs will be restored. an archive contains large objects, then all large objects will be restored.
</para> </para>
</listitem> </listitem>
...@@ -620,8 +620,8 @@ connectDBStart() -- connect() failed: No such file or directory ...@@ -620,8 +620,8 @@ connectDBStart() -- connect() failed: No such file or directory
</para> </para>
<para> <para>
To dump a database called mydb that contains To dump a database called <literal>mydb</> that contains
<acronym>BLOB</acronym>s to a <filename>tar</filename> file: large objects to a <filename>tar</filename> file:
<screen> <screen>
<prompt>$</prompt> <userinput>pg_dump -Ft -b mydb &gt; db.tar</userinput> <prompt>$</prompt> <userinput>pg_dump -Ft -b mydb &gt; db.tar</userinput>
...@@ -629,8 +629,8 @@ connectDBStart() -- connect() failed: No such file or directory ...@@ -629,8 +629,8 @@ connectDBStart() -- connect() failed: No such file or directory
</para> </para>
<para> <para>
To reload this database (with <acronym>BLOB</acronym>s) to an To reload this database (with large objects) to an
existing database called newdb: existing database called <literal>newdb</>:
<screen> <screen>
<prompt>$</prompt> <userinput>pg_restore -d newdb db.tar</userinput> <prompt>$</prompt> <userinput>pg_restore -d newdb db.tar</userinput>
......
...@@ -15,7 +15,7 @@ ...@@ -15,7 +15,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_archiver.c,v 1.32 2001/08/22 20:23:23 petere Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_archiver.c,v 1.33 2001/09/21 21:58:30 petere Exp $
* *
* Modifications - 28-Jun-2000 - pjw@rhyme.com.au * Modifications - 28-Jun-2000 - pjw@rhyme.com.au
* *
...@@ -333,7 +333,7 @@ RestoreArchive(Archive *AHX, RestoreOptions *ropt) ...@@ -333,7 +333,7 @@ RestoreArchive(Archive *AHX, RestoreOptions *ropt)
* warnings. * warnings.
*/ */
if (!AH->CustomOutPtr) if (!AH->CustomOutPtr)
write_msg(modulename, "WARNING: skipping BLOB restoration\n"); write_msg(modulename, "WARNING: skipping large object restoration\n");
} }
else else
...@@ -398,12 +398,12 @@ RestoreArchive(Archive *AHX, RestoreOptions *ropt) ...@@ -398,12 +398,12 @@ RestoreArchive(Archive *AHX, RestoreOptions *ropt)
if ((reqs & 2) != 0) /* We loaded the data */ if ((reqs & 2) != 0) /* We loaded the data */
{ {
ahlog(AH, 1, "fixing up BLOB reference for %s\n", te->name); ahlog(AH, 1, "fixing up large object cross-reference for %s\n", te->name);
FixupBlobRefs(AH, te->name); FixupBlobRefs(AH, te->name);
} }
} }
else else
ahlog(AH, 2, "ignoring BLOB cross-references for %s %s\n", te->desc, te->name); ahlog(AH, 2, "ignoring large object cross-references for %s %s\n", te->desc, te->name);
te = te->next; te = te->next;
} }
...@@ -717,7 +717,7 @@ StartBlob(Archive *AHX, Oid oid) ...@@ -717,7 +717,7 @@ StartBlob(Archive *AHX, Oid oid)
ArchiveHandle *AH = (ArchiveHandle *) AHX; ArchiveHandle *AH = (ArchiveHandle *) AHX;
if (!AH->StartBlobPtr) if (!AH->StartBlobPtr)
die_horribly(AH, modulename, "BLOB output not supported in chosen format\n"); die_horribly(AH, modulename, "large object output not supported in chosen format\n");
(*AH->StartBlobPtr) (AH, AH->currToc, oid); (*AH->StartBlobPtr) (AH, AH->currToc, oid);
...@@ -757,14 +757,14 @@ EndRestoreBlobs(ArchiveHandle *AH) ...@@ -757,14 +757,14 @@ EndRestoreBlobs(ArchiveHandle *AH)
{ {
if (AH->txActive) if (AH->txActive)
{ {
ahlog(AH, 2, "committing BLOB transactions\n"); ahlog(AH, 2, "committing large object transactions\n");
CommitTransaction(AH); CommitTransaction(AH);
} }
if (AH->blobTxActive) if (AH->blobTxActive)
CommitTransactionXref(AH); CommitTransactionXref(AH);
ahlog(AH, 1, "restored %d BLOBs\n", AH->blobCount); ahlog(AH, 1, "restored %d large objects\n", AH->blobCount);
} }
...@@ -781,7 +781,7 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid) ...@@ -781,7 +781,7 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid)
if (!AH->createdBlobXref) if (!AH->createdBlobXref)
{ {
if (!AH->connection) if (!AH->connection)
die_horribly(AH, modulename, "cannot restore BLOBs without a database connection\n"); die_horribly(AH, modulename, "cannot restore large objects without a database connection\n");
CreateBlobXrefTable(AH); CreateBlobXrefTable(AH);
AH->createdBlobXref = 1; AH->createdBlobXref = 1;
...@@ -792,7 +792,7 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid) ...@@ -792,7 +792,7 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid)
*/ */
if (!AH->txActive) if (!AH->txActive)
{ {
ahlog(AH, 2, "starting BLOB transactions\n"); ahlog(AH, 2, "starting large object transactions\n");
StartTransaction(AH); StartTransaction(AH);
} }
if (!AH->blobTxActive) if (!AH->blobTxActive)
...@@ -800,15 +800,15 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid) ...@@ -800,15 +800,15 @@ StartRestoreBlob(ArchiveHandle *AH, Oid oid)
loOid = lo_creat(AH->connection, INV_READ | INV_WRITE); loOid = lo_creat(AH->connection, INV_READ | INV_WRITE);
if (loOid == 0) if (loOid == 0)
die_horribly(AH, modulename, "could not create BLOB\n"); die_horribly(AH, modulename, "could not create large object\n");
ahlog(AH, 2, "restoring BLOB oid %u as %u\n", oid, loOid); ahlog(AH, 2, "restoring large object with oid %u as %u\n", oid, loOid);
InsertBlobXref(AH, oid, loOid); InsertBlobXref(AH, oid, loOid);
AH->loFd = lo_open(AH->connection, loOid, INV_WRITE); AH->loFd = lo_open(AH->connection, loOid, INV_WRITE);
if (AH->loFd == -1) if (AH->loFd == -1)
die_horribly(AH, modulename, "could not open BLOB\n"); die_horribly(AH, modulename, "could not open large object\n");
AH->writingBlob = 1; AH->writingBlob = 1;
} }
...@@ -824,7 +824,7 @@ EndRestoreBlob(ArchiveHandle *AH, Oid oid) ...@@ -824,7 +824,7 @@ EndRestoreBlob(ArchiveHandle *AH, Oid oid)
*/ */
if (((AH->blobCount / BLOB_BATCH_SIZE) * BLOB_BATCH_SIZE) == AH->blobCount) if (((AH->blobCount / BLOB_BATCH_SIZE) * BLOB_BATCH_SIZE) == AH->blobCount)
{ {
ahlog(AH, 2, "committing BLOB transactions\n"); ahlog(AH, 2, "committing large object transactions\n");
CommitTransaction(AH); CommitTransaction(AH);
CommitTransactionXref(AH); CommitTransactionXref(AH);
} }
...@@ -1198,7 +1198,7 @@ ahwrite(const void *ptr, size_t size, size_t nmemb, ArchiveHandle *AH) ...@@ -1198,7 +1198,7 @@ ahwrite(const void *ptr, size_t size, size_t nmemb, ArchiveHandle *AH)
if (AH->writingBlob) if (AH->writingBlob)
{ {
res = lo_write(AH->connection, AH->loFd, (void *) ptr, size * nmemb); res = lo_write(AH->connection, AH->loFd, (void *) ptr, size * nmemb);
ahlog(AH, 5, "wrote %d bytes of BLOB data (result = %d)\n", size * nmemb, res); ahlog(AH, 5, "wrote %d bytes of large object data (result = %d)\n", size * nmemb, res);
if (res < size * nmemb) if (res < size * nmemb)
die_horribly(AH, modulename, "could not write to large object (result: %d, expected: %d)\n", die_horribly(AH, modulename, "could not write to large object (result: %d, expected: %d)\n",
res, size * nmemb); res, size * nmemb);
......
...@@ -19,7 +19,7 @@ ...@@ -19,7 +19,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_custom.c,v 1.14 2001/08/19 22:17:03 petere Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_custom.c,v 1.15 2001/09/21 21:58:30 petere Exp $
* *
* Modifications - 28-Jun-2000 - pjw@rhyme.com.au * Modifications - 28-Jun-2000 - pjw@rhyme.com.au
* *
...@@ -386,7 +386,7 @@ static void ...@@ -386,7 +386,7 @@ static void
_StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid)
{ {
if (oid == 0) if (oid == 0)
die_horribly(AH, modulename, "invalid OID for BLOB\n"); die_horribly(AH, modulename, "invalid OID for large object\n");
WriteInt(AH, oid); WriteInt(AH, oid);
_StartDataCompressor(AH, te); _StartDataCompressor(AH, te);
...@@ -503,7 +503,7 @@ _PrintTocData(ArchiveHandle *AH, TocEntry *te, RestoreOptions *ropt) ...@@ -503,7 +503,7 @@ _PrintTocData(ArchiveHandle *AH, TocEntry *te, RestoreOptions *ropt)
case BLK_BLOBS: case BLK_BLOBS:
if (!AH->connection) if (!AH->connection)
die_horribly(AH, modulename, "BLOBs cannot be loaded without a database connection\n"); die_horribly(AH, modulename, "large objects cannot be loaded without a database connection\n");
_LoadBlobs(AH); _LoadBlobs(AH);
break; break;
......
...@@ -5,7 +5,7 @@ ...@@ -5,7 +5,7 @@
* Implements the basic DB functions used by the archiver. * Implements the basic DB functions used by the archiver.
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_db.c,v 1.25 2001/09/17 02:07:51 inoue Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_db.c,v 1.26 2001/09/21 21:58:30 petere Exp $
* *
* NOTES * NOTES
* *
...@@ -738,7 +738,7 @@ FixupBlobRefs(ArchiveHandle *AH, char *tablename) ...@@ -738,7 +738,7 @@ FixupBlobRefs(ArchiveHandle *AH, char *tablename)
{ {
attr = PQgetvalue(res, i, 0); attr = PQgetvalue(res, i, 0);
ahlog(AH, 1, "fixing BLOB cross-references for %s.%s\n", tablename, attr); ahlog(AH, 1, "fixing large object cross-references for %s.%s\n", tablename, attr);
resetPQExpBuffer(tblQry); resetPQExpBuffer(tblQry);
...@@ -785,16 +785,16 @@ CreateBlobXrefTable(ArchiveHandle *AH) ...@@ -785,16 +785,16 @@ CreateBlobXrefTable(ArchiveHandle *AH)
if (!AH->blobConnection) if (!AH->blobConnection)
AH->blobConnection = _connectDB(AH, NULL, NULL); AH->blobConnection = _connectDB(AH, NULL, NULL);
ahlog(AH, 1, "creating table for BLOB cross-references\n"); ahlog(AH, 1, "creating table for large object cross-references\n");
appendPQExpBuffer(qry, "Create Temporary Table %s(oldOid oid, newOid oid);", BLOB_XREF_TABLE); appendPQExpBuffer(qry, "Create Temporary Table %s(oldOid oid, newOid oid);", BLOB_XREF_TABLE);
ExecuteSqlCommand(AH, qry, "could not create BLOB cross reference table", true); ExecuteSqlCommand(AH, qry, "could not create large object cross-reference table", true);
resetPQExpBuffer(qry); resetPQExpBuffer(qry);
appendPQExpBuffer(qry, "Create Unique Index %s_ix on %s(oldOid)", BLOB_XREF_TABLE, BLOB_XREF_TABLE); appendPQExpBuffer(qry, "Create Unique Index %s_ix on %s(oldOid)", BLOB_XREF_TABLE, BLOB_XREF_TABLE);
ExecuteSqlCommand(AH, qry, "could not create index on BLOB cross reference table", true); ExecuteSqlCommand(AH, qry, "could not create index on large object cross-reference table", true);
destroyPQExpBuffer(qry); destroyPQExpBuffer(qry);
} }
...@@ -806,7 +806,7 @@ InsertBlobXref(ArchiveHandle *AH, int old, int new) ...@@ -806,7 +806,7 @@ InsertBlobXref(ArchiveHandle *AH, int old, int new)
appendPQExpBuffer(qry, "Insert Into %s(oldOid, newOid) Values (%d, %d);", BLOB_XREF_TABLE, old, new); appendPQExpBuffer(qry, "Insert Into %s(oldOid, newOid) Values (%d, %d);", BLOB_XREF_TABLE, old, new);
ExecuteSqlCommand(AH, qry, "could not create BLOB cross reference entry", true); ExecuteSqlCommand(AH, qry, "could not create large object cross-reference entry", true);
destroyPQExpBuffer(qry); destroyPQExpBuffer(qry);
} }
...@@ -832,7 +832,7 @@ StartTransactionXref(ArchiveHandle *AH) ...@@ -832,7 +832,7 @@ StartTransactionXref(ArchiveHandle *AH)
appendPQExpBuffer(qry, "Begin;"); appendPQExpBuffer(qry, "Begin;");
ExecuteSqlCommand(AH, qry, ExecuteSqlCommand(AH, qry,
"could not start transaction for BLOB cross references", true); "could not start transaction for large object cross-references", true);
AH->blobTxActive = true; AH->blobTxActive = true;
destroyPQExpBuffer(qry); destroyPQExpBuffer(qry);
...@@ -858,7 +858,7 @@ CommitTransactionXref(ArchiveHandle *AH) ...@@ -858,7 +858,7 @@ CommitTransactionXref(ArchiveHandle *AH)
appendPQExpBuffer(qry, "Commit;"); appendPQExpBuffer(qry, "Commit;");
ExecuteSqlCommand(AH, qry, "could not commit transaction for BLOB cross references", true); ExecuteSqlCommand(AH, qry, "could not commit transaction for large object cross-references", true);
AH->blobTxActive = false; AH->blobTxActive = false;
destroyPQExpBuffer(qry); destroyPQExpBuffer(qry);
......
...@@ -20,7 +20,7 @@ ...@@ -20,7 +20,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_files.c,v 1.12 2001/07/03 20:21:48 petere Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_files.c,v 1.13 2001/09/21 21:58:30 petere Exp $
* *
* Modifications - 28-Jun-2000 - pjw@rhyme.com.au * Modifications - 28-Jun-2000 - pjw@rhyme.com.au
* *
...@@ -363,7 +363,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt) ...@@ -363,7 +363,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt)
ctx->blobToc = fopen("blobs.toc", PG_BINARY_R); ctx->blobToc = fopen("blobs.toc", PG_BINARY_R);
if (ctx->blobToc == NULL) if (ctx->blobToc == NULL)
die_horribly(AH, modulename, "could not open BLOB TOC for input: %s\n", strerror(errno)); die_horribly(AH, modulename, "could not open large object TOC for input: %s\n", strerror(errno));
_getBlobTocEntry(AH, &oid, fname); _getBlobTocEntry(AH, &oid, fname);
...@@ -376,7 +376,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt) ...@@ -376,7 +376,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt)
} }
if (fclose(ctx->blobToc) != 0) if (fclose(ctx->blobToc) != 0)
die_horribly(AH, modulename, "could not close BLOB TOC file: %s\n", strerror(errno)); die_horribly(AH, modulename, "could not close large object TOC file: %s\n", strerror(errno));
EndRestoreBlobs(AH); EndRestoreBlobs(AH);
} }
...@@ -474,7 +474,7 @@ _StartBlobs(ArchiveHandle *AH, TocEntry *te) ...@@ -474,7 +474,7 @@ _StartBlobs(ArchiveHandle *AH, TocEntry *te)
if (ctx->blobToc == NULL) if (ctx->blobToc == NULL)
die_horribly(AH, modulename, die_horribly(AH, modulename,
"could not open BLOB TOC for output: %s\n", strerror(errno)); "could not open large object TOC for output: %s\n", strerror(errno));
} }
...@@ -495,7 +495,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) ...@@ -495,7 +495,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid)
char *sfx; char *sfx;
if (oid == 0) if (oid == 0)
die_horribly(AH, modulename, "invalid OID for BLOB (%u)\n", oid); die_horribly(AH, modulename, "invalid OID for large object (%u)\n", oid);
if (AH->compression != 0) if (AH->compression != 0)
sfx = ".gz"; sfx = ".gz";
...@@ -514,7 +514,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) ...@@ -514,7 +514,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid)
#endif #endif
if (tctx->FH == NULL) if (tctx->FH == NULL)
die_horribly(AH, modulename, "could not open BLOB file\n"); die_horribly(AH, modulename, "could not open large object file\n");
} }
/* /*
...@@ -529,7 +529,7 @@ _EndBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) ...@@ -529,7 +529,7 @@ _EndBlob(ArchiveHandle *AH, TocEntry *te, Oid oid)
lclTocEntry *tctx = (lclTocEntry *) te->formatData; lclTocEntry *tctx = (lclTocEntry *) te->formatData;
if (GZCLOSE(tctx->FH) != 0) if (GZCLOSE(tctx->FH) != 0)
die_horribly(AH, modulename, "could not close BLOB file\n"); die_horribly(AH, modulename, "could not close large object file\n");
} }
/* /*
...@@ -547,7 +547,7 @@ _EndBlobs(ArchiveHandle *AH, TocEntry *te) ...@@ -547,7 +547,7 @@ _EndBlobs(ArchiveHandle *AH, TocEntry *te)
/* WriteInt(AH, 0); */ /* WriteInt(AH, 0); */
if (fclose(ctx->blobToc) != 0) if (fclose(ctx->blobToc) != 0)
die_horribly(AH, modulename, "could not close BLOB TOC file: %s\n", strerror(errno)); die_horribly(AH, modulename, "could not close large object TOC file: %s\n", strerror(errno));
} }
...@@ -16,7 +16,7 @@ ...@@ -16,7 +16,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_tar.c,v 1.17 2001/07/03 20:21:48 petere Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_backup_tar.c,v 1.18 2001/09/21 21:58:30 petere Exp $
* *
* Modifications - 28-Jun-2000 - pjw@rhyme.com.au * Modifications - 28-Jun-2000 - pjw@rhyme.com.au
* *
...@@ -715,7 +715,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt) ...@@ -715,7 +715,7 @@ _LoadBlobs(ArchiveHandle *AH, RestoreOptions *ropt)
if (strncmp(th->targetFile, "blob_", 5) == 0 && oid != 0) if (strncmp(th->targetFile, "blob_", 5) == 0 && oid != 0)
{ {
ahlog(AH, 1, "restoring BLOB oid %u\n", oid); ahlog(AH, 1, "restoring large object OID %u\n", oid);
StartRestoreBlob(AH, oid); StartRestoreBlob(AH, oid);
...@@ -911,7 +911,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid) ...@@ -911,7 +911,7 @@ _StartBlob(ArchiveHandle *AH, TocEntry *te, Oid oid)
char *sfx; char *sfx;
if (oid == 0) if (oid == 0)
die_horribly(AH, modulename, "invalid OID for BLOB (%u)\n", oid); die_horribly(AH, modulename, "invalid OID for large object (%u)\n", oid);
if (AH->compression != 0) if (AH->compression != 0)
sfx = ".gz"; sfx = ".gz";
......
...@@ -22,7 +22,7 @@ ...@@ -22,7 +22,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_dump.c,v 1.229 2001/09/07 01:11:50 tgl Exp $ * $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_dump.c,v 1.230 2001/09/21 21:58:30 petere Exp $
* *
*------------------------------------------------------------------------- *-------------------------------------------------------------------------
*/ */
...@@ -141,7 +141,7 @@ help(const char *progname) ...@@ -141,7 +141,7 @@ help(const char *progname)
#ifdef HAVE_GETOPT_LONG #ifdef HAVE_GETOPT_LONG
puts(gettext( puts(gettext(
" -a, --data-only dump only the data, not the schema\n" " -a, --data-only dump only the data, not the schema\n"
" -b, --blobs include BLOB data in dump\n" " -b, --blobs include large objects in dump\n"
" -c, --clean clean (drop) schema prior to create\n" " -c, --clean clean (drop) schema prior to create\n"
" -C, --create include commands to create database in dump\n" " -C, --create include commands to create database in dump\n"
" -d, --inserts dump data as INSERT, rather than COPY, commands\n" " -d, --inserts dump data as INSERT, rather than COPY, commands\n"
...@@ -175,7 +175,7 @@ help(const char *progname) ...@@ -175,7 +175,7 @@ help(const char *progname)
#else #else
puts(gettext( puts(gettext(
" -a dump only the data, not the schema\n" " -a dump only the data, not the schema\n"
" -b include BLOB data in dump\n" " -b include large objects in dump\n"
" -c clean (drop) schema prior to create\n" " -c clean (drop) schema prior to create\n"
" -C include commands to create database in dump\n" " -C include commands to create database in dump\n"
" -d dump data as INSERT, rather than COPY, commands\n" " -d dump data as INSERT, rather than COPY, commands\n"
...@@ -961,7 +961,7 @@ main(int argc, char **argv) ...@@ -961,7 +961,7 @@ main(int argc, char **argv)
if (outputBlobs && tablename != NULL && strlen(tablename) > 0) if (outputBlobs && tablename != NULL && strlen(tablename) > 0)
{ {
write_msg(NULL, "BLOB output is not supported for a single table.\n"); write_msg(NULL, "Large object output is not supported for a single table.\n");
write_msg(NULL, "Use all tables or a full dump instead.\n"); write_msg(NULL, "Use all tables or a full dump instead.\n");
exit(1); exit(1);
} }
...@@ -975,7 +975,7 @@ main(int argc, char **argv) ...@@ -975,7 +975,7 @@ main(int argc, char **argv)
if (outputBlobs == true && (format[0] == 'p' || format[0] == 'P')) if (outputBlobs == true && (format[0] == 'p' || format[0] == 'P'))
{ {
write_msg(NULL, "BLOB output is not supported for plain text dump files.\n"); write_msg(NULL, "large object output is not supported for plain text dump files.\n");
write_msg(NULL, "(Use a different output format.)\n"); write_msg(NULL, "(Use a different output format.)\n");
exit(1); exit(1);
} }
...@@ -1216,7 +1216,7 @@ dumpBlobs(Archive *AH, char *junkOid, void *junkVal) ...@@ -1216,7 +1216,7 @@ dumpBlobs(Archive *AH, char *junkOid, void *junkVal)
Oid blobOid; Oid blobOid;
if (g_verbose) if (g_verbose)
write_msg(NULL, "saving BLOBs\n"); write_msg(NULL, "saving large objects\n");
/* Cursor to get all BLOB tables */ /* Cursor to get all BLOB tables */
if (AH->remoteVersion >= 70100) if (AH->remoteVersion >= 70100)
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment