-
+
Data Types
string.
- If one explicitly casts a value to character
- varying(n>) or
- character(n>) , then an over-length
- value will be truncated to n> characters without
- raising an error. (This too is required by the
-
-
-
- Prior to
PostgreSQL> 7.2, strings that were too long were
- always truncated without raising an error, in either explicit or
- implicit casting contexts.
-
-
+ If one explicitly casts a value to character
+ varying(n>) or
+ character(n>) , then an over-length
+ value will be truncated to n> characters without
+ raising an error. (This too is required by the
+
The notations varchar(n>) and
-
- Prior to
PostgreSQL> 7.2, bit data
- was always silently truncated or zero-padded on the right, with
- or without an explicit cast. This was changed to comply with the
-
-
-
Refer to
linkend="sql-syntax-bit-strings"> for information about the syntax
-
+
Date/Time Support
Gregorian years AD 1-99 may be entered by using 4 digits with leading
- zeros (e.g., 0099> is AD 99). Previous versions of
-
PostgreSQL accepted years with three
- digits and with single digits, but as of version 7.0 the rules have
- been tightened up to reduce the possibility of ambiguity.
+ zeros (e.g., 0099> is AD 99).
-
+
Data Definition
Deprecated
- In
previous versions of PostgreSQL , the
+ In
releases of PostgreSQL prior to 7.1 , the
default behavior was not to include child tables in queries. This was
- found to be error prone and is also in violation of the SQL
- standard. Under the old syntax, to include the child tables you append
- * to the table name. For example:
-SELECT * from cities*;
-
- You can still explicitly specify scanning child tables by
- appending * , as well as explicitly specify not
- scanning child tables by writing ONLY . But
- beginning in version 7.1, the default behavior for an undecorated
- table name is to scan its child tables too, whereas before the
- default was not to do so. To get the old default behavior,
- disable the configuration
+ found to be error prone and also in violation of the SQL
+ standard. You can get the pre-7.1 behavior by turning off the
+ configuration
option.
-
+
Functions and Operators
the result is given to the full available precision.
-
- Prior to
PostgreSQL 7.2, the precision
- parameters were unimplemented, and the result was always given
- in integer seconds.
-
-
-
Some examples:
-
+
libpq application programmers should be careful to
maintain the PGconn abstraction. Use the accessor
-functions described below to get
-at the contents of PGconn . Avoid directly referencing the fields of the
-PGconn> structure because they are subject to change in the future.
-(Beginning in
PostgreSQL release 6.4, the
-definition of the struct behind PGconn> is not even provided in libpq-fe.h .
-If you have old code that accesses PGconn fields directly, you can keep using it
-by including libpq-int.h too, but you are encouraged to fix the code
-soon.)
+functions described below to get at the contents of PGconn .
+Reference to internal PGconn fields using
+libpq-int.h> is not recommended because they are subject to change
+in the future.
typedef struct pgNotify {
char *relname; /* notification condition name */
- int be_pid; /* process ID of server process */
+ int be_pid; /* process ID of notifying server process */
char *extra; /* notification parameter */
} PGnotify;
always point to an empty string.)
-
- In
PostgreSQL 6.4 and later,
- the be_pid is that of the notifying server process,
- whereas in earlier versions it was always the
PID of your own server process.
-
-
-
gives a sample program that illustrates the use
of asynchronous notification.
- If your codes references the header file
- libpq-int.h and you refuse to fix your code to
- not use it, starting in
PostgreSQL> 7.2, this file will be found in
- includedir /postgresql/internal/libpq-int.h ,
- so you need to add the appropriate -I option to
- your compiler command line.
-
-
-
+
Large Objects
values. This is not described here.
- history">
-
History
+ intro">
+
Introduction
-
POSTGRES 4.2 , the indirect predecessor
- of
PostgreSQL , supported three standard
- implementations of large objects: as files external to the
-
POSTGRES server, as external files
- managed by the
POSTGRES server, and as
- data stored within the
POSTGRES
- database. This caused considerable confusion among users. As a
- result, only support for large objects as data stored within the
- database is retained in
PostgreSQL .
- Even though this is slower to access, it provides stricter data
- integrity. For historical reasons, this storage scheme is
- referred to as Inversion large
- objects. (You will see the term Inversion used
- occasionally to mean the same thing as large object.) Since
-
PostgreSQL 7.1 , all large objects are
- placed in one system table called
- pg_largeobject .
-
+
+ versus large objects
+
-
- versus large objects
-
-
PostgreSQL 7.1 introduced a mechanism
- (nicknamed
TOAST
) that allows
- data values to be much larger than single pages. This
- makes the large object facility partially obsolete. One
+ All large objects are placed in a single system table called
+ pg_largeobject .
+
PostgreSQL also supports a storage system called
+
TOAST
that automatically stores values
+ larger than a single database page into a secondary storage area per table.
+ This makes the large object facility partially obsolete. One
remaining advantage of the large object facility is that it allows values
up to 2 GB in size, whereas
TOAST ed fields can be at
- most 1 GB. Also, large objects can be manipulated piece-by-piece much more
- easily than ordinary data fields, so the practical limits are considerably
- different .
+ most 1 GB. Also, large objects can be randomly modified using a read/write
+ API that is more efficient than performing such operations using
Implementation Features
- The large object implementation breaks large
- objects up into chunks
and stores the chunks in
+ The large object implementation breaks large
+ objects up into chunks
and stores the chunks in
rows in the database. A B-tree index guarantees fast
searches for the correct chunk number when doing random
access reads and writes.
PostgreSQL client interface libraries
provide for accessing large objects. All large object
manipulation using these functions must take
- place within an SQL transaction block. (This requirement is
- strictly enforced as of
PostgreSQL 6.5>, though it
- has been an implicit requirement in previous versions, resulting
- in misbehavior if ignored.)
+ place within an SQL transaction block.
The
PostgreSQL large object interface is modeled after
the
Unix file-system interface, with analogues of
open , read ,
-
+
Routine Database Maintenance Tasks
- Beginning in
PostgreSQL 7.2, the standard form
- of VACUUM> can run in parallel with normal database operations
- (selects, inserts, updates, deletes, but not changes to table definitions).
- Routine vacuuming is therefore not nearly as intrusive as it was in prior
- releases, and it is not as critical to try to schedule it at low-usage
- times of day.
-
-
+ The standard form of VACUUM> can run in parallel with
+ normal database operations (SELECTs, INSERTs, UPDATEs, DELETEs, but not
+ changes to table definitions).
Beginning in
PostgreSQL 8.0, there are
configuration parameters that can be adjusted to further reduce the
performance impact of background vacuuming. See
It is possible to run ANALYZE> on specific tables and even
just specific columns of a table, so the flexibility exists to update some
statistics more frequently than others if your application requires it.
- In practice, however, the usefulness of this feature is doubtful.
- Beginning in
PostgreSQL 7.2,
- ANALYZE> is a fairly fast operation even on large tables,
- because it uses a statistical random sampling of the rows of a table
- rather than reading every single row. So it's probably much simpler
- to just run it over the whole database every so often.
+ In practice, however, it is usually best to just analyze the entire database
+ because it is a fast operation. It uses a statistical random sampling of
+ the rows of a table rather than reading every single row.
transactions that were in the past appear to be in the future — which
means their outputs become invisible. In short, catastrophic data loss.
(Actually the data is still there, but that's cold comfort if you can't
- get at it.)
-
-
- Prior to
PostgreSQL 7.2, the only defense
- against XID wraparound was to re-initdb> at least every 4
- billion transactions. This of course was not very satisfactory for
- high-traffic sites, so a better solution has been devised. The new
- approach allows a server to remain up indefinitely, without
- initdb> or any sort of restart. The price is this
- maintenance requirement: every table in the database must
- be vacuumed at least once every billion transactions.
+ get at it.) To avoid this, it is necessary to vacuum every table
+ in every database at least once every billion transactions.
-
+
Concurrency Control
TABLE locks the whole table.) This should be taken into
account when porting applications to
PostgreSQL from other environments.
- (Before version 6.5
PostgreSQL used
- read locks, and so this above consideration is also relevant when
- upgrading from
PostgreSQL versions
- prior to 6.5.)
-
+
The Rule System
Another situation is cases on UPDATE where it depends on the
change of an attribute if an action should be performed or
- not. In
PostgreSQL version 6.4, the
- attribute specification for rule events is disabled (it will have
- its comeback latest in 6.5, maybe earlier
- - stay tuned). So for now the only way to
+ not. The only way to
create a rule as in the shoelace_log example is to do it with
a rule qualification. That results in an extra query that is
performed always, even if the attribute of interest cannot
-
+
-Since
PostgreSQL uses a fixed page size (commonly
-8Kb), and does not allow tuples to span multiple pages, it's not possible to
-store very large field values directly. Before
PostgreSQL> 7.1
-there was a hard limit of just under one page on the total amount of data that
-could be put into a table row. In release 7.1 and later, this limit is
-overcome by allowing large field values to be compressed and/or broken up into
-multiple physical rows. This happens transparently to the user, with only
-small impact on most of the backend code. The technique is affectionately
-known as
TOAST> (or the best thing since sliced bread>).
+
PostgreSQL uses a fixed page size (commonly
+8Kb), and does not allow tuples to span multiple pages. Therefore, it is
+not possible to store very large field values directly. To overcome
+this limitation, large field values are compressed and/or broken up into
+multiple physical rows. This happens transparently to the user, with only
+small impact on most of the backend code. The technique is affectionately
+known as
TOAST> (or the best thing since sliced bread>).
-
+
User-Defined Functions
command pg_config --pkglibdir .
- Before
PostgreSQL release 7.2, only
- exact absolute paths to object files could be specified in
- CREATE FUNCTION>. This approach is now deprecated
- since it makes the function definition unnecessarily unportable.
- It's best to specify just the shared library name with no path nor
- extension, and let the search mechanism provide that information
- instead.
-
--includedir-server
pg_config>with user-defined C functions>>
to find out where the
PostgreSQL> server header
files are installed on your system (or the system that your
- users will be running on). This option is new with
-
PostgreSQL> 7.1 you should use the option
- --includedir . (pg_config
- will exit with a non-zero status if it encounters an unknown
- option.) For releases prior to 7.1 you will have to guess,
- but since that was before the current calling conventions were
- introduced, it is unlikely that you want to support those
- releases.
+ users will be running on).