mirror of
https://github.com/psycopg/psycopg2.git
synced 2024-12-02 05:33:45 +03:00
2cde9033ac
Not implemented yet!
639 lines
23 KiB
ReStructuredText
639 lines
23 KiB
ReStructuredText
Basic module usage
|
|
==================
|
|
|
|
.. sectionauthor:: Daniele Varrazzo <daniele.varrazzo@gmail.com>
|
|
|
|
.. index::
|
|
pair: Example; Usage
|
|
|
|
The basic Psycopg usage is common to all the database adapters implementing
|
|
the |DBAPI|_ protocol. Here is an interactive session showing some of the
|
|
basic commands::
|
|
|
|
>>> import psycopg2
|
|
|
|
# Connect to an existing database
|
|
>>> conn = psycopg2.connect("dbname=test user=postgres")
|
|
|
|
# Open a cursor to perform database operations
|
|
>>> cur = conn.cursor()
|
|
|
|
# Execute a command: this creates a new table
|
|
>>> cur.execute("CREATE TABLE test (id serial PRIMARY KEY, num integer, data varchar);")
|
|
|
|
# Pass data to fill a query placeholders and let Psycopg perform
|
|
# the correct conversion (no more SQL injections!)
|
|
>>> cur.execute("INSERT INTO test (num, data) VALUES (%s, %s)",
|
|
... (100, "abc'def"))
|
|
|
|
# Query the database and obtain data as Python objects
|
|
>>> cur.execute("SELECT * FROM test;")
|
|
>>> cur.fetchone()
|
|
(1, 100, "abc'def")
|
|
|
|
# Make the changes to the database persistent
|
|
>>> conn.commit()
|
|
|
|
# Close communication with the database
|
|
>>> cur.close()
|
|
>>> conn.close()
|
|
|
|
|
|
The main entry point of Psycopg are:
|
|
|
|
- The function `~psycopg2.connect()` creates a new database session and
|
|
returns a new `connection` instance.
|
|
|
|
- The class `connection` encapsulates a database session. It allows to:
|
|
|
|
- create new `cursor`\s using the `~connection.cursor()` method to
|
|
execute database commands and queries,
|
|
|
|
- terminate the session using the methods `~connection.commit()` or
|
|
`~connection.rollback()`.
|
|
|
|
- The class `cursor` allows interaction with the database:
|
|
|
|
- send commands to the database using methods such as `~cursor.execute()`
|
|
and `~cursor.executemany()`,
|
|
|
|
- retrieve data from the database :ref:`by iteration <cursor-iterable>` or
|
|
using methods such as `~cursor.fetchone()`, `~cursor.fetchmany()`,
|
|
`~cursor.fetchall()`.
|
|
|
|
|
|
|
|
.. index::
|
|
pair: Query; Parameters
|
|
|
|
.. _query-parameters:
|
|
|
|
Passing parameters to SQL queries
|
|
---------------------------------
|
|
|
|
Psycopg casts Python variables to SQL literals by type. Many standard Python types
|
|
are already `adapted to the correct SQL representation`__.
|
|
|
|
.. __: python-types-adaptation_
|
|
|
|
Example: the Python function call::
|
|
|
|
>>> cur.execute(
|
|
... """INSERT INTO some_table (an_int, a_date, a_string)
|
|
... VALUES (%s, %s, %s);""",
|
|
... (10, datetime.date(2005, 11, 18), "O'Reilly"))
|
|
|
|
is converted into the SQL command::
|
|
|
|
INSERT INTO some_table (an_int, a_date, a_string)
|
|
VALUES (10, '2005-11-18', 'O''Reilly');
|
|
|
|
Named arguments are supported too using :samp:`%({name})s` placeholders.
|
|
Using named arguments the values can be passed to the query in any order and
|
|
many placeholder can use the same values::
|
|
|
|
>>> cur.execute(
|
|
... """INSERT INTO some_table (an_int, a_date, another_date, a_string)
|
|
... VALUES (%(int)s, %(date)s, %(date)s, %(str)s);""",
|
|
... {'int': 10, 'str': "O'Reilly", 'date': datetime.date(2005, 11, 18)})
|
|
|
|
While the mechanism resembles regular Python strings manipulation, there are a
|
|
few subtle differences you should care about when passing parameters to a
|
|
query:
|
|
|
|
- The Python string operator ``%`` is not used: the `~cursor.execute()`
|
|
method accepts a tuple or dictionary of values as second parameter.
|
|
|sql-warn|__.
|
|
|
|
.. |sql-warn| replace:: **Never** use ``%`` or ``+`` to merge values
|
|
into queries
|
|
|
|
.. __: sql-injection_
|
|
|
|
- The variables placeholder must *always be a* ``%s``, even if a different
|
|
placeholder (such as a ``%d`` for integers or ``%f`` for floats) may look
|
|
more appropriate::
|
|
|
|
>>> cur.execute("INSERT INTO numbers VALUES (%d)", (42,)) # WRONG
|
|
>>> cur.execute("INSERT INTO numbers VALUES (%s)", (42,)) # correct
|
|
|
|
- For positional variables binding, *the second argument must always be a
|
|
sequence*, even if it contains a single variable. And remember that Python
|
|
requires a comma to create a single element tuple::
|
|
|
|
>>> cur.execute("INSERT INTO foo VALUES (%s)", "bar") # WRONG
|
|
>>> cur.execute("INSERT INTO foo VALUES (%s)", ("bar")) # WRONG
|
|
>>> cur.execute("INSERT INTO foo VALUES (%s)", ("bar",)) # correct
|
|
>>> cur.execute("INSERT INTO foo VALUES (%s)", ["bar"]) # correct
|
|
|
|
- Only variable values should be bound via this method: it shouldn't be used
|
|
to set table or field names. For these elements, ordinary string formatting
|
|
should be used before running `~cursor.execute()`.
|
|
|
|
|
|
|
|
.. index:: Security, SQL injection
|
|
|
|
.. _sql-injection:
|
|
|
|
The problem with the query parameters
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
The SQL representation for many data types is often not the same of the Python
|
|
string representation. The classic example is with single quotes in
|
|
strings: SQL uses them as string constants bounds and requires them to be
|
|
escaped, whereas in Python single quotes can be left unescaped in strings
|
|
bounded by double quotes. For this reason a naïve approach to the composition
|
|
of query strings, e.g. using string concatenation, is a recipe for terrible
|
|
problems::
|
|
|
|
>>> SQL = "INSERT INTO authors (name) VALUES ('%s');" # NEVER DO THIS
|
|
>>> data = ("O'Reilly", )
|
|
>>> cur.execute(SQL % data) # THIS WILL FAIL MISERABLY
|
|
ProgrammingError: syntax error at or near "Reilly"
|
|
LINE 1: INSERT INTO authors (name) VALUES ('O'Reilly')
|
|
^
|
|
|
|
If the variable containing the data to be sent to the database comes from an
|
|
untrusted source (e.g. a form published on a web site) an attacker could
|
|
easily craft a malformed string, either gaining access to unauthorized data or
|
|
performing destructive operations on the database. This form of attack is
|
|
called `SQL injection`_ and is known to be one of the most widespread forms of
|
|
attack to servers. Before continuing, please print `this page`__ as a memo and
|
|
hang it onto your desk.
|
|
|
|
.. _SQL injection: http://en.wikipedia.org/wiki/SQL_injection
|
|
.. __: http://xkcd.com/327/
|
|
|
|
Psycopg can `convert automatically Python objects into and from SQL
|
|
literals`__: using this feature your code will result more robust and
|
|
reliable. It is really the case to stress this point:
|
|
|
|
.. __: python-types-adaptation_
|
|
|
|
.. warning::
|
|
|
|
Never, **never**, **NEVER** use Python string concatenation (``+``) or
|
|
string parameters interpolation (``%``) to pass variables to a SQL query
|
|
string. Not even at gunpoint.
|
|
|
|
The correct way to pass variables in a SQL command is using the second
|
|
argument of the `~cursor.execute()` method::
|
|
|
|
>>> SQL = "INSERT INTO authors (name) VALUES (%s);" # Notice: no quotes
|
|
>>> data = ("O'Reilly", )
|
|
>>> cur.execute(SQL, data) # Notice: no % operator
|
|
|
|
|
|
|
|
.. index::
|
|
single: Adaptation
|
|
pair: Objects; Adaptation
|
|
single: Data types; Adaptation
|
|
|
|
.. _python-types-adaptation:
|
|
|
|
Adaptation of Python values to SQL types
|
|
----------------------------------------
|
|
|
|
Many standards Python types are adapted into SQL and returned as Python
|
|
objects when a query is executed.
|
|
|
|
If you need to convert other Python types to and from PostgreSQL data types,
|
|
see :ref:`adapting-new-types` and :ref:`type-casting-from-sql-to-python`. You
|
|
can also find a few other specialized adapters in the `psycopg2.extras`
|
|
module.
|
|
|
|
In the following examples the method `~cursor.mogrify()` is used to show
|
|
the SQL string that would be sent to the database.
|
|
|
|
.. index::
|
|
pair: None; Adaptation
|
|
single: NULL; Adaptation
|
|
pair: Boolean; Adaptation
|
|
|
|
- Python ``None`` and boolean values are converted into the proper SQL
|
|
literals::
|
|
|
|
>>> cur.mogrify("SELECT %s, %s, %s;", (None, True, False))
|
|
>>> 'SELECT NULL, true, false;'
|
|
|
|
.. index::
|
|
single: Adaptation; numbers
|
|
single: Integer; Adaptation
|
|
single: Float; Adaptation
|
|
single: Decimal; Adaptation
|
|
|
|
- Numeric objects: `!int`, `!long`, `!float`,
|
|
`!Decimal` are converted in the PostgreSQL numerical representation::
|
|
|
|
>>> cur.mogrify("SELECT %s, %s, %s, %s;", (10, 10L, 10.0, Decimal("10.00")))
|
|
>>> 'SELECT 10, 10, 10.0, 10.00;'
|
|
|
|
.. index::
|
|
pair: Strings; Adaptation
|
|
single: Unicode; Adaptation
|
|
single: Buffer; Adaptation
|
|
single: bytea; Adaptation
|
|
single: Binary string
|
|
|
|
- String types: `!str`, `!unicode` are converted in SQL string
|
|
syntax. `!buffer` is converted in PostgreSQL binary string syntax,
|
|
suitable for :sql:`bytea` fields. When reading textual fields, either
|
|
`!str` or `!unicode` can be received: see
|
|
:ref:`unicode-handling`.
|
|
|
|
.. index::
|
|
single: Adaptation; Date/Time objects
|
|
single: Date objects; Adaptation
|
|
single: Time objects; Adaptation
|
|
single: Interval objects; Adaptation
|
|
single: mx.DateTime; Adaptation
|
|
|
|
- Date and time objects: builtin `!datetime`, `!date`,
|
|
`!time`. `!timedelta` are converted into PostgreSQL's
|
|
:sql:`timestamp`, :sql:`date`, :sql:`time`, :sql:`interval` data types.
|
|
Time zones are supported too. The Egenix `mx.DateTime`_ objects are adapted
|
|
the same way::
|
|
|
|
>>> dt = datetime.datetime.now()
|
|
>>> dt
|
|
datetime.datetime(2010, 2, 8, 1, 40, 27, 425337)
|
|
|
|
>>> cur.mogrify("SELECT %s, %s, %s;", (dt, dt.date(), dt.time()))
|
|
"SELECT '2010-02-08T01:40:27.425337', '2010-02-08', '01:40:27.425337';"
|
|
|
|
>>> cur.mogrify("SELECT %s;", (dt - datetime.datetime(2010,1,1),))
|
|
"SELECT '38 days 6027.425337 seconds';"
|
|
|
|
.. index::
|
|
single: Array; Adaptation
|
|
double: Lists; Adaptation
|
|
|
|
- Python lists are converted into PostgreSQL :sql:`ARRAY`\ s::
|
|
|
|
>>> cur.mogrify("SELECT %s;", ([10, 20, 30], ))
|
|
'SELECT ARRAY[10, 20, 30];'
|
|
|
|
.. index::
|
|
double: Tuple; Adaptation
|
|
single: IN operator
|
|
|
|
- Python tuples are converted in a syntax suitable for the SQL :sql:`IN`
|
|
operator and to represent a composite type::
|
|
|
|
>>> cur.mogrify("SELECT %s IN %s;", (10, (10, 20, 30)))
|
|
'SELECT 10 IN (10, 20, 30);'
|
|
|
|
.. note::
|
|
|
|
SQL doesn't allow an empty list in the IN operator, so your code should
|
|
guard against empty tuples.
|
|
|
|
If you want PostgreSQL composite types to be converted into a Python
|
|
tuple/namedtuple you can use the `~psycopg2.extras.register_composite()`
|
|
function.
|
|
|
|
.. versionadded:: 2.0.6
|
|
the tuple :sql:`IN` adaptation.
|
|
|
|
.. versionchanged:: 2.0.14
|
|
the tuple :sql:`IN` adapter is always active. In previous releases it
|
|
was necessary to import the `~psycopg2.extensions` module to have it
|
|
registered.
|
|
|
|
.. versionchanged:: 2.3
|
|
named tuples are adapted like regular tuples and can thus be used to
|
|
represent composite types.
|
|
|
|
- Python dictionaries are converted into the |hstore|_ data type. See
|
|
`~psycopg2.extras.register_hstore()` for further details.
|
|
|
|
.. |hstore| replace:: :sql:`hstore`
|
|
.. _hstore: http://www.postgresql.org/docs/9.0/static/hstore.html
|
|
|
|
.. versionadded:: 2.3
|
|
the :sql:`hstore` adaptation.
|
|
|
|
|
|
.. index::
|
|
single: Unicode
|
|
|
|
.. _unicode-handling:
|
|
|
|
Unicode handling
|
|
^^^^^^^^^^^^^^^^
|
|
|
|
Psycopg can exchange Unicode data with a PostgreSQL database. Python
|
|
`!unicode` objects are automatically *encoded* in the client encoding
|
|
defined on the database connection (the `PostgreSQL encoding`__, available in
|
|
`connection.encoding`, is translated into a `Python codec`__ using the
|
|
`~psycopg2.extensions.encodings` mapping)::
|
|
|
|
>>> print u, type(u)
|
|
àèìòù€ <type 'unicode'>
|
|
|
|
>>> cur.execute("INSERT INTO test (num, data) VALUES (%s,%s);", (74, u))
|
|
|
|
.. __: http://www.postgresql.org/docs/9.0/static/multibyte.html
|
|
.. __: http://docs.python.org/library/codecs.html#standard-encodings
|
|
|
|
When reading data from the database, the strings returned are usually 8 bit
|
|
`!str` objects encoded in the database client encoding::
|
|
|
|
>>> print conn.encoding
|
|
UTF8
|
|
|
|
>>> cur.execute("SELECT data FROM test WHERE num = 74")
|
|
>>> x = cur.fetchone()[0]
|
|
>>> print x, type(x), repr(x)
|
|
àèìòù€ <type 'str'> '\xc3\xa0\xc3\xa8\xc3\xac\xc3\xb2\xc3\xb9\xe2\x82\xac'
|
|
|
|
>>> conn.set_client_encoding('LATIN9')
|
|
|
|
>>> cur.execute("SELECT data FROM test WHERE num = 74")
|
|
>>> x = cur.fetchone()[0]
|
|
>>> print type(x), repr(x)
|
|
<type 'str'> '\xe0\xe8\xec\xf2\xf9\xa4'
|
|
|
|
In order to obtain `!unicode` objects instead, it is possible to
|
|
register a typecaster so that PostgreSQL textual types are automatically
|
|
*decoded* using the current client encoding::
|
|
|
|
>>> psycopg2.extensions.register_type(psycopg2.extensions.UNICODE, cur)
|
|
|
|
>>> cur.execute("SELECT data FROM test WHERE num = 74")
|
|
>>> x = cur.fetchone()[0]
|
|
>>> print x, type(x), repr(x)
|
|
àèìòù€ <type 'unicode'> u'\xe0\xe8\xec\xf2\xf9\u20ac'
|
|
|
|
In the above example, the `~psycopg2.extensions.UNICODE` typecaster is
|
|
registered only on the cursor. It is also possible to register typecasters on
|
|
the connection or globally: see the function
|
|
`~psycopg2.extensions.register_type()` and
|
|
:ref:`type-casting-from-sql-to-python` for details.
|
|
|
|
.. note::
|
|
|
|
If you want to receive uniformly all your database input in Unicode, you
|
|
can register the related typecasters globally as soon as Psycopg is
|
|
imported::
|
|
|
|
import psycopg2
|
|
import psycopg2.extensions
|
|
psycopg2.extensions.register_type(psycopg2.extensions.UNICODE)
|
|
psycopg2.extensions.register_type(psycopg2.extensions.UNICODEARRAY)
|
|
|
|
and then forget about this story.
|
|
|
|
|
|
.. index::
|
|
single: Time Zones
|
|
|
|
.. _tz-handling:
|
|
|
|
Time zones handling
|
|
^^^^^^^^^^^^^^^^^^^
|
|
|
|
The PostgreSQL type :sql:`timestamp with time zone` is converted into Python
|
|
`!datetime` objects with a `!tzinfo` attribute set to a
|
|
`~psycopg2.tz.FixedOffsetTimezone` instance.
|
|
|
|
>>> cur.execute("SET TIME ZONE 'Europe/Rome';") # UTC + 1 hour
|
|
>>> cur.execute("SELECT '2010-01-01 10:30:45'::timestamptz;")
|
|
>>> cur.fetchone()[0].tzinfo
|
|
psycopg2.tz.FixedOffsetTimezone(offset=60, name=None)
|
|
|
|
Notice that only time zones with an integer number of minutes are supported:
|
|
this is a limitation of the Python `!datetime` module. A few historical time
|
|
zones had seconds in the UTC offset: these time zones will have the offset
|
|
rounded to the nearest minute, with an error of up to 30 seconds.
|
|
|
|
>>> cur.execute("SET TIME ZONE 'Asia/Calcutta';") # offset was +5:53:20
|
|
>>> cur.execute("SELECT '1930-01-01 10:30:45'::timestamptz;")
|
|
>>> cur.fetchone()[0].tzinfo
|
|
psycopg2.tz.FixedOffsetTimezone(offset=353, name=None)
|
|
|
|
.. versionchanged:: 2.2.2
|
|
timezones with seconds are supported (with rounding). Previously such
|
|
timezones raised an error. In order to deal with them in previous
|
|
versions use `psycopg2.extras.register_tstz_w_secs`.
|
|
|
|
|
|
.. index:: Transaction, Begin, Commit, Rollback, Autocommit
|
|
|
|
.. _transactions-control:
|
|
|
|
Transactions control
|
|
--------------------
|
|
|
|
In Psycopg transactions are handled by the `connection` class. By
|
|
default, the first time a command is sent to the database (using one of the
|
|
`cursor`\ s created by the connection), a new transaction is created.
|
|
The following database commands will be executed in the context of the same
|
|
transaction -- not only the commands issued by the first cursor, but the ones
|
|
issued by all the cursors created by the same connection. Should any command
|
|
fail, the transaction will be aborted and no further command will be executed
|
|
until a call to the `connection.rollback()` method.
|
|
|
|
The connection is responsible to terminate its transaction, calling either the
|
|
`~connection.commit()` or `~connection.rollback()` method. Committed
|
|
changes are immediately made persistent into the database. Closing the
|
|
connection using the `~connection.close()` method or destroying the
|
|
connection object (calling `!__del__()` or letting it fall out of scope)
|
|
will result in an implicit `!rollback()` call.
|
|
|
|
It is possible to set the connection in *autocommit* mode: this way all the
|
|
commands executed will be immediately committed and no rollback is possible. A
|
|
few commands (e.g. :sql:`CREATE DATABASE`, :sql:`VACUUM`...) require to be run
|
|
outside any transaction: in order to be able to run these commands from
|
|
Psycopg, the session must be in autocommit mode. Read the documentation for
|
|
`connection.set_isolation_level()` to know how to change the commit mode.
|
|
|
|
|
|
|
|
.. index::
|
|
pair: Server side; Cursor
|
|
pair: Named; Cursor
|
|
pair: DECLARE; SQL command
|
|
pair: FETCH; SQL command
|
|
pair: MOVE; SQL command
|
|
|
|
.. _server-side-cursors:
|
|
|
|
Server side cursors
|
|
-------------------
|
|
|
|
When a database query is executed, the Psycopg `cursor` usually fetches
|
|
all the records returned by the backend, transferring them to the client
|
|
process. If the query returned an huge amount of data, a proportionally large
|
|
amount of memory will be allocated by the client.
|
|
|
|
If the dataset is too large to be practically handled on the client side, it is
|
|
possible to create a *server side* cursor. Using this kind of cursor it is
|
|
possible to transfer to the client only a controlled amount of data, so that a
|
|
large dataset can be examined without keeping it entirely in memory.
|
|
|
|
Server side cursor are created in PostgreSQL using the |DECLARE|_ command and
|
|
subsequently handled using :sql:`MOVE`, :sql:`FETCH` and :sql:`CLOSE` commands.
|
|
|
|
Psycopg wraps the database server side cursor in *named cursors*. A named
|
|
cursor is created using the `~connection.cursor()` method specifying the
|
|
`name` parameter. Such cursor will behave mostly like a regular cursor,
|
|
allowing the user to move in the dataset using the `~cursor.scroll()`
|
|
method and to read the data using `~cursor.fetchone()` and
|
|
`~cursor.fetchmany()` methods.
|
|
|
|
.. |DECLARE| replace:: :sql:`DECLARE`
|
|
.. _DECLARE: http://www.postgresql.org/docs/9.0/static/sql-declare.html
|
|
|
|
|
|
|
|
.. index:: Thread safety, Multithread, Multiprocess
|
|
|
|
.. _thread-safety:
|
|
|
|
Thread and process safety
|
|
-------------------------
|
|
|
|
The Psycopg module and the `connection` objects are *thread-safe*: many
|
|
threads can access the same database either using separate sessions and
|
|
creating a `!connection` per thread or using the same using the same
|
|
connection and creating separate `cursor`\ s. In |DBAPI|_ parlance, Psycopg is
|
|
*level 2 thread safe*.
|
|
|
|
The difference between the above two approaches is that, using different
|
|
connections, the commands will be executed in different sessions and will be
|
|
served by different server processes. On the other hand, using many cursors on
|
|
the same connection, all the commands will be executed in the same session
|
|
(and in the same transaction if the connection is not in :ref:`autocommit
|
|
<transactions-control>` mode), but they will be serialized.
|
|
|
|
The above observations are only valid for regular threads: they don't apply to
|
|
forked processes nor to green threads. `libpq` connections `shouldn't be used by a
|
|
forked processes`__, so when using a module such as |multiprocessing|__ or a
|
|
forking web deploy method such as FastCGI ensure to create the connections
|
|
*after* the fork.
|
|
|
|
.. __: http://www.postgresql.org/docs/9.0/static/libpq-connect.html#LIBPQ-CONNECT
|
|
.. |multiprocessing| replace:: `!multiprocessing`
|
|
.. __: http://docs.python.org/library/multiprocessing.html
|
|
|
|
Connections shouldn't be shared either by different green threads: doing so
|
|
may result in a deadlock. See :ref:`green-support` for further details.
|
|
|
|
|
|
|
|
.. index::
|
|
pair: COPY; SQL command
|
|
|
|
.. _copy:
|
|
|
|
Using COPY TO and COPY FROM
|
|
---------------------------
|
|
|
|
Psycopg `cursor` objects provide an interface to the efficient
|
|
PostgreSQL |COPY|__ command to move data from files to tables and back.
|
|
The methods exposed are:
|
|
|
|
`~cursor.copy_from()`
|
|
Reads data *from* a file-like object appending them to a database table
|
|
(:sql:`COPY table FROM file` syntax). The source file must have both
|
|
`!read()` and `!readline()` method.
|
|
|
|
`~cursor.copy_to()`
|
|
Writes the content of a table *to* a file-like object (:sql:`COPY table TO
|
|
file` syntax). The target file must have a `write()` method.
|
|
|
|
`~cursor.copy_expert()`
|
|
Allows to handle more specific cases and to use all the :sql:`COPY`
|
|
features available in PostgreSQL.
|
|
|
|
Please refer to the documentation of the single methods for details and
|
|
examples.
|
|
|
|
.. |COPY| replace:: :sql:`COPY`
|
|
.. __: http://www.postgresql.org/docs/9.0/static/sql-copy.html
|
|
|
|
|
|
|
|
.. index::
|
|
single: Large objects
|
|
|
|
.. _large-objects:
|
|
|
|
Access to PostgreSQL large objects
|
|
----------------------------------
|
|
|
|
PostgreSQL offers support to `large objects`__, which provide stream-style
|
|
access to user data that is stored in a special large-object structure. They
|
|
are useful with data values too large to be manipulated conveniently as a
|
|
whole.
|
|
|
|
.. __: http://www.postgresql.org/docs/9.0/static/largeobjects.html
|
|
|
|
Psycopg allows access to the large object using the
|
|
`~psycopg2.extensions.lobject` class. Objects are generated using the
|
|
`connection.lobject()` factory method. Data can be retrieved either as bytes
|
|
or as Unicode strings.
|
|
|
|
Psycopg large object support efficient import/export with file system files
|
|
using the |lo_import|_ and |lo_export|_ libpq functions.
|
|
|
|
.. |lo_import| replace:: `!lo_import()`
|
|
.. _lo_import: http://www.postgresql.org/docs/9.0/static/lo-interfaces.html#LO-IMPORT
|
|
.. |lo_export| replace:: `!lo_export()`
|
|
.. _lo_export: http://www.postgresql.org/docs/9.0/static/lo-interfaces.html#LO-EXPORT
|
|
|
|
|
|
|
|
.. index::
|
|
pair: Two-phase commit; Transaction
|
|
|
|
.. _tpc:
|
|
|
|
Two-Phase Commit protocol support
|
|
---------------------------------
|
|
|
|
.. versionadded:: 2.3
|
|
|
|
Psycopg exposes the two-phase commit features available since PostgreSQL 8.1
|
|
implementing the *two-phase commit extensions* proposed by the |DBAPI|.
|
|
|
|
The |DBAPI| model of two-phase commit is inspired to the `XA specification`__,
|
|
according to which transaction IDs are formed from three components:
|
|
|
|
- a format ID (non-negative 32 bit integer)
|
|
- a global transaction ID (string not longer than 64 bytes)
|
|
- a branch qualifier (string not longer than 64 bytes)
|
|
|
|
For a particular global transaction, the first two components will be the same
|
|
for all the resources. Every resource will be assigned a different branch
|
|
qualifier.
|
|
|
|
According to the |DBAPI| specification, a transaction ID is created using the
|
|
`connection.xid()` method. Once you have a transaction id, a distributed
|
|
transaction can be started with `connection.tpc_begin()`, prepared using
|
|
`~connection.tpc_prepare()` and completed using `~connection.tpc_commit()` or
|
|
`~connection.tpc_rollback()`. Transaction IDs can also be retrieved from the
|
|
database using `~connection.tpc_recover()` and completed using the above
|
|
`!tpc_commit()` and `!tpc_rollback()`.
|
|
|
|
PostgreSQL doesn't follow the XA standard though, and the ID for a PostgreSQL
|
|
prepared transaction can be any string up to 200 characters long.
|
|
Psycopg's `~psycopg2.extensions.Xid` objects can represent both XA-style
|
|
transactions IDs (such as the ones created by the `!xid()` method) and
|
|
PostgreSQL transaction IDs identified by an unparsed string.
|
|
|
|
The format in which the Xids are converted into strings passed to the
|
|
database is the same employed by the `PostgreSQL JDBC driver`__: this should
|
|
allow interoperation between tools written in Python and in Java. For example
|
|
a recovery tool written in Python would be able to recognize the components of
|
|
transactions produced by a Java program.
|
|
|
|
For further details see the documentation for the above methods.
|
|
|
|
.. __: http://www.opengroup.org/bookstore/catalog/c193.htm
|
|
.. __: http://jdbc.postgresql.org/
|
|
|