diff --git a/AUTHORS b/AUTHORS index 32017591..60677ddb 100644 --- a/AUTHORS +++ b/AUTHORS @@ -6,7 +6,7 @@ For the win32 port: Jason Erickson Additional Help: - + Peter Fein contributed a logging connection/cursor class that even if it was not used directly heavily influenced the implementation currently in psycopg2.extras. diff --git a/LICENSE b/LICENSE index dd7c124d..360a44f6 100644 --- a/LICENSE +++ b/LICENSE @@ -47,8 +47,8 @@ psycopg/microprotocol*.{h,c}: claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required. - + 2. Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software. - + 3. This notice may not be removed or altered from any source distribution. diff --git a/doc/COPYING.LESSER b/doc/COPYING.LESSER index fc8a5de7..cca7fc27 100644 --- a/doc/COPYING.LESSER +++ b/doc/COPYING.LESSER @@ -10,7 +10,7 @@ the terms and conditions of version 3 of the GNU General Public License, supplemented by the additional permissions listed below. - 0. Additional Definitions. + 0. Additional Definitions. As used herein, "this License" refers to version 3 of the GNU Lesser General Public License, and the "GNU GPL" refers to version 3 of the GNU @@ -111,7 +111,7 @@ the following: a copy of the Library already present on the user's computer system, and (b) will operate properly with a modified version of the Library that is interface-compatible with the Linked - Version. + Version. e) Provide Installation Information, but only if you would otherwise be required to provide such information under section 6 of the diff --git a/doc/SUCCESS b/doc/SUCCESS index 7788b13f..de459918 100644 --- a/doc/SUCCESS +++ b/doc/SUCCESS @@ -1,10 +1,10 @@ From: Jack Moffitt To: Psycopg Mailing List Subject: Re: [Psycopg] preparing for 1.0 -Date: 22 Oct 2001 11:16:21 -0600 +Date: 22 Oct 2001 11:16:21 -0600 -www.vorbis.com is serving from 5-10k pages per day with psycopg serving -data for most of that. +www.vorbis.com is serving from 5-10k pages per day with psycopg serving +data for most of that. I plan to use it for several of our other sites, so that number will increase. @@ -19,7 +19,7 @@ jack. From: Yury Don To: Psycopg Mailing List Subject: Re: [Psycopg] preparing for 1.0 -Date: 23 Oct 2001 09:53:11 +0600 +Date: 23 Oct 2001 09:53:11 +0600 We use psycopg and psycopg zope adapter since fisrt public release (it seems version 0.4). Now it works on 3 our sites and in intranet @@ -32,7 +32,7 @@ to solve the problem, even thouth my knowledge of c were poor. BTW, segfault with dictfetchall on particular data set (see [Psycopg] dictfetchXXX() problems) disappeared in 0.99.8pre2. --- +-- Best regards, Yury Don @@ -42,7 +42,7 @@ To: Federico Di Gregorio Cc: Psycopg Mailing List Subject: Re: [Psycopg] preparing for 1.0 Date: 23 Oct 2001 08:25:52 -0400 - + The US Govt Department of Labor's Office of Disability Employment Policy's DisabilityDirect website is run on zope and zpsycopg. @@ -50,7 +50,7 @@ Policy's DisabilityDirect website is run on zope and zpsycopg. From: Scott Leerssen To: Federico Di Gregorio Subject: Re: [Psycopg] preparing for 1.0 -Date: 23 Oct 2001 09:56:10 -0400 +Date: 23 Oct 2001 09:56:10 -0400 Racemi's load management software infrastructure uses psycopg to handle complex server allocation decisions, plus storage and access of @@ -66,10 +66,10 @@ From: Andre Schubert To: Federico Di Gregorio Cc: Psycopg Mailing List Subject: Re: [Psycopg] preparing for 1.0 -Date: 23 Oct 2001 11:46:07 +0200 +Date: 23 Oct 2001 11:46:07 +0200 i have changed the psycopg version to 0.99.8pre2 on all devel-machines -and all segfaults are gone. after my holiday i wil change to 0.99.8pre2 +and all segfaults are gone. after my holiday i wil change to 0.99.8pre2 or 1.0 on our production-server. this server contains several web-sites which are all connected to postgres over ZPsycopgDA. @@ -81,7 +81,7 @@ From: Fred Wilson Horch To: Subject: [Psycopg] Success story for psycopg Date: 23 Oct 2001 10:59:17 -0400 - + Due to various quirks of PyGreSQL and PoPy, EcoAccess has been looking for a reliable, fast and relatively bug-free Python-PostgreSQL interface for our project. @@ -98,7 +98,7 @@ reports and feature requests, and we're looking forward to using psycopg as the Python interface for additional database-backed web applications. Keep up the good work! --- +-- Fred Wilson Horch mailto:fhorch@ecoaccess.org Executive Director, EcoAccess http://ecoaccess.org/ diff --git a/doc/pep-0249.txt b/doc/pep-0249.txt index e0e36002..e74fd0db 100644 --- a/doc/pep-0249.txt +++ b/doc/pep-0249.txt @@ -9,15 +9,15 @@ Replaces: 248 Release-Date: 07 Apr 1999 Introduction - + This API has been defined to encourage similarity between the Python modules that are used to access databases. By doing this, we hope to achieve a consistency leading to more easily understood modules, code that is generally more portable across databases, and a broader reach of database connectivity from Python. - + The interface specification consists of several sections: - + * Module Interface * Connection Objects * Cursor Objects @@ -25,7 +25,7 @@ Introduction * Type Objects and Constructors * Implementation Hints * Major Changes from 1.0 to 2.0 - + Comments and questions about this specification may be directed to the SIG for Database Interfacing with Python (db-sig@python.org). @@ -41,7 +41,7 @@ Introduction basis for new interfaces. Module Interface - + Access to the database is made available through connection objects. The module must provide the following constructor for these: @@ -51,17 +51,17 @@ Module Interface Constructor for creating a connection to the database. Returns a Connection Object. It takes a number of parameters which are database dependent. [1] - + These module globals must be defined: apilevel String constant stating the supported DB API level. Currently only the strings '1.0' and '2.0' are allowed. - + If not given, a DB-API 1.0 level interface should be assumed. - + threadsafety Integer constant stating the level of thread safety the @@ -81,33 +81,33 @@ Module Interface or other external sources that are beyond your control. paramstyle - + String constant stating the type of parameter marker formatting expected by the interface. Possible values are [2]: - 'qmark' Question mark style, + 'qmark' Question mark style, e.g. '...WHERE name=?' - 'numeric' Numeric, positional style, + 'numeric' Numeric, positional style, e.g. '...WHERE name=:1' - 'named' Named style, + 'named' Named style, e.g. '...WHERE name=:name' - 'format' ANSI C printf format codes, + 'format' ANSI C printf format codes, e.g. '...WHERE name=%s' - 'pyformat' Python extended format codes, + 'pyformat' Python extended format codes, e.g. '...WHERE name=%(name)s' The module should make all error information available through these exceptions or subclasses thereof: - Warning - + Warning + Exception raised for important warnings like data truncations while inserting, etc. It must be a subclass of the Python StandardError (defined in the module exceptions). - - Error + + Error Exception that is the base class of all other error exceptions. You can use this to catch all errors with one @@ -115,7 +115,7 @@ Module Interface errors and thus should not use this class as base. It must be a subclass of the Python StandardError (defined in the module exceptions). - + InterfaceError Exception raised for errors that are related to the @@ -126,50 +126,50 @@ Module Interface Exception raised for errors that are related to the database. It must be a subclass of Error. - + DataError - + Exception raised for errors that are due to problems with the processed data like division by zero, numeric value out of range, etc. It must be a subclass of DatabaseError. - + OperationalError - + Exception raised for errors that are related to the database's operation and not necessarily under the control of the programmer, e.g. an unexpected disconnect occurs, the data source name is not found, a transaction could not be processed, a memory allocation error occurred during processing, etc. It must be a subclass of DatabaseError. - - IntegrityError - + + IntegrityError + Exception raised when the relational integrity of the database is affected, e.g. a foreign key check fails. It must be a subclass of DatabaseError. - - InternalError - + + InternalError + Exception raised when the database encounters an internal error, e.g. the cursor is not valid anymore, the transaction is out of sync, etc. It must be a subclass of DatabaseError. - + ProgrammingError - + Exception raised for programming errors, e.g. table not found or already exists, syntax error in the SQL statement, wrong number of parameters specified, etc. It must be a subclass of DatabaseError. - + NotSupportedError - + Exception raised in case a method or database API was used which is not supported by the database, e.g. requesting a .rollback() on a connection that does not support transaction or has transactions turned off. It must be a subclass of DatabaseError. - + This is the exception inheritance layout: StandardError @@ -183,17 +183,17 @@ Module Interface |__InternalError |__ProgrammingError |__NotSupportedError - + Note: The values of these exceptions are not defined. They should give the user a fairly good idea of what went wrong, though. - + Connection Objects Connection Objects should respond to the following methods: - .close() - + .close() + Close the connection now (rather than whenever __del__ is called). The connection will be unusable from this point forward; an Error (or subclass) exception will be raised @@ -203,52 +203,52 @@ Connection Objects committing the changes first will cause an implicit rollback to be performed. - + .commit() - + Commit any pending transaction to the database. Note that if the database supports an auto-commit feature, this must be initially off. An interface method may be provided to turn it back on. - + Database modules that do not support transactions should implement this method with void functionality. - - .rollback() - + + .rollback() + This method is optional since not all databases provide transaction support. [3] - + In case a database does provide transactions this method causes the the database to roll back to the start of any pending transaction. Closing a connection without committing the changes first will cause an implicit rollback to be performed. - + .cursor() - + Return a new Cursor Object using the connection. If the database does not provide a direct cursor concept, the module will have to emulate cursors using other means to the extent needed by this specification. [4] - + Cursor Objects These objects represent a database cursor, which is used to - manage the context of a fetch operation. Cursors created from + manage the context of a fetch operation. Cursors created from the same connection are not isolated, i.e., any changes done to the database by a cursor are immediately visible by the other cursors. Cursors created from different connections can or can not be isolated, depending on how the transaction support - is implemented (see also the connection's rollback() and commit() + is implemented (see also the connection's rollback() and commit() methods.) - + Cursor Objects should respond to the following methods and attributes: - .description - + .description + This read-only attribute is a sequence of 7-item sequences. Each of these sequences contains information describing one result column: (name, type_code, @@ -260,17 +260,17 @@ Cursor Objects This attribute will be None for operations that do not return rows or if the cursor has not had an operation invoked via the executeXXX() method yet. - + The type_code can be interpreted by comparing it to the Type Objects specified in the section below. - - .rowcount - + + .rowcount + This read-only attribute specifies the number of rows that the last executeXXX() produced (for DQL statements like 'select') or affected (for DML statements like 'update' or 'insert'). - + The attribute is -1 in case no executeXXX() has been performed on the cursor or the rowcount of the last operation is not determinable by the interface. [7] @@ -278,96 +278,96 @@ Cursor Objects Note: Future versions of the DB API specification could redefine the latter case to have the object return None instead of -1. - + .callproc(procname[,parameters]) - + (This method is optional since not all databases provide stored procedures. [3]) - + Call a stored database procedure with the given name. The sequence of parameters must contain one entry for each argument that the procedure expects. The result of the call is returned as modified copy of the input sequence. Input parameters are left untouched, output and input/output parameters replaced with possibly new values. - + The procedure may also provide a result set as output. This must then be made available through the standard fetchXXX() methods. - + .close() - + Close the cursor now (rather than whenever __del__ is called). The cursor will be unusable from this point forward; an Error (or subclass) exception will be raised if any operation is attempted with the cursor. - - .execute(operation[,parameters]) - + + .execute(operation[,parameters]) + Prepare and execute a database operation (query or command). Parameters may be provided as sequence or mapping and will be bound to variables in the operation. Variables are specified in a database-specific notation (see the module's paramstyle attribute for details). [5] - + A reference to the operation will be retained by the cursor. If the same operation object is passed in again, then the cursor can optimize its behavior. This is most effective for algorithms where the same operation is used, but different parameters are bound to it (many times). - + For maximum efficiency when reusing an operation, it is best to use the setinputsizes() method to specify the parameter types and sizes ahead of time. It is legal for a parameter to not match the predefined information; the implementation should compensate, possibly with a loss of efficiency. - + The parameters may also be specified as list of tuples to e.g. insert multiple rows in a single operation, but this kind of usage is depreciated: executemany() should be used instead. - + Return values are not defined. - - .executemany(operation,seq_of_parameters) - + + .executemany(operation,seq_of_parameters) + Prepare a database operation (query or command) and then execute it against all parameter sequences or mappings found in the sequence seq_of_parameters. - + Modules are free to implement this method using multiple calls to the execute() method or by using array operations to have the database process the sequence as a whole in one call. - + Use of this method for an operation which produces one or more result sets constitutes undefined behavior, and the - implementation is permitted (but not required) to raise + implementation is permitted (but not required) to raise an exception when it detects that a result set has been created by an invocation of the operation. - + The same comments as for execute() also apply accordingly to this method. - + Return values are not defined. - - .fetchone() - + + .fetchone() + Fetch the next row of a query result set, returning a single sequence, or None when no more data is available. [6] - + An Error (or subclass) exception is raised if the previous call to executeXXX() did not produce any result set or no call was issued yet. fetchmany([size=cursor.arraysize]) - + Fetch the next set of rows of a query result, returning a sequence of sequences (e.g. a list of tuples). An empty sequence is returned when no more rows are available. - + The number of rows to fetch per call is specified by the parameter. If it is not given, the cursor's arraysize determines the number of rows to be fetched. The method @@ -375,62 +375,62 @@ Cursor Objects parameter. If this is not possible due to the specified number of rows not being available, fewer rows may be returned. - + An Error (or subclass) exception is raised if the previous call to executeXXX() did not produce any result set or no call was issued yet. - + Note there are performance considerations involved with the size parameter. For optimal performance, it is usually best to use the arraysize attribute. If the size parameter is used, then it is best for it to retain the same value from one fetchmany() call to the next. - - .fetchall() + + .fetchall() Fetch all (remaining) rows of a query result, returning them as a sequence of sequences (e.g. a list of tuples). Note that the cursor's arraysize attribute can affect the performance of this operation. - + An Error (or subclass) exception is raised if the previous call to executeXXX() did not produce any result set or no call was issued yet. - - .nextset() - + + .nextset() + (This method is optional since not all databases support multiple result sets. [3]) - + This method will make the cursor skip to the next available set, discarding any remaining rows from the current set. - + If there are no more sets, the method returns None. Otherwise, it returns a true value and subsequent calls to the fetch methods will return rows from the next result set. - + An Error (or subclass) exception is raised if the previous call to executeXXX() did not produce any result set or no call was issued yet. .arraysize - + This read/write attribute specifies the number of rows to fetch at a time with fetchmany(). It defaults to 1 meaning to fetch a single row at a time. - + Implementations must observe this value with respect to the fetchmany() method, but are free to interact with the database a single row at a time. It may also be used in the implementation of executemany(). - + .setinputsizes(sizes) - + This can be used before a call to executeXXX() to predefine memory areas for the operation's parameters. - + sizes is specified as a sequence -- one item for each input parameter. The item should be a Type Object that corresponds to the input that will be used, or it should @@ -438,27 +438,27 @@ Cursor Objects parameter. If the item is None, then no predefined memory area will be reserved for that column (this is useful to avoid predefined areas for large inputs). - + This method would be used before the executeXXX() method is invoked. - + Implementations are free to have this method do nothing and users are free to not use it. - + .setoutputsize(size[,column]) - + Set a column buffer size for fetches of large columns (e.g. LONGs, BLOBs, etc.). The column is specified as an index into the result sequence. Not specifying the column will set the default size for all large columns in the cursor. - + This method would be used before the executeXXX() method is invoked. - + Implementations are free to have this method do nothing and users are free to not use it. - + Type Objects and Constructors @@ -485,15 +485,15 @@ Type Objects and Constructors Implementation Hints below for details). The module exports the following constructors and singletons: - + Date(year,month,day) This function constructs an object holding a date value. - + Time(hour,minute,second) This function constructs an object holding a time value. - + Timestamp(year,month,day,hour,minute,second) This function constructs an object holding a time stamp @@ -507,12 +507,12 @@ Type Objects and Constructors module for details). TimeFromTicks(ticks) - + This function constructs an object holding a time value from the given ticks value (number of seconds since the epoch; see the documentation of the standard Python time module for details). - + TimestampFromTicks(ticks) This function constructs an object holding a time stamp @@ -521,10 +521,10 @@ Type Objects and Constructors time module for details). Binary(string) - + This function constructs an object capable of holding a binary (long) string value. - + STRING @@ -535,22 +535,22 @@ Type Objects and Constructors This type object is used to describe (long) binary columns in a database (e.g. LONG, RAW, BLOBs). - + NUMBER This type object is used to describe numeric columns in a database. DATETIME - + This type object is used to describe date/time columns in a database. - + ROWID - + This type object is used to describe the "Row ID" column in a database. - + SQL NULL values are represented by the Python None singleton on input and output. @@ -563,7 +563,7 @@ Implementation Hints for Module Authors * The preferred object types for the date/time objects are those defined in the mxDateTime package. It provides all necessary constructors and methods both at Python and C level. - + * The preferred object type for Binary objects are the buffer types available in standard Python starting with version 1.5.2. Please see the Python documentation for @@ -577,7 +577,7 @@ Implementation Hints for Module Authors processing. However, it should be noted that this does not expose a C API like mxDateTime does which means that integration with C based database modules is more difficult. - + * Here is a sample implementation of the Unix ticks based constructors for date/time delegating work to the generic constructors: @@ -645,7 +645,7 @@ Implementation Hints for Module Authors class NotSupportedError(DatabaseError): pass - + In C you can use the PyErr_NewException(fullname, base, NULL) API to create the exception objects. @@ -760,7 +760,7 @@ Optional DB API Extensions Warning Message: "DB-API extension connection.messages used" Cursor Method .next() - + Return the next row from the currently executing SQL statement using the same semantics as .fetchone(). A StopIteration exception is raised when the result set is exhausted for Python @@ -790,13 +790,13 @@ Optional DB API Extensions Warning Message: "DB-API extension cursor.lastrowid used" - + Optional Error Handling Extension The core DB API specification only introduces a set of exceptions which can be raised to report errors to the user. In some cases, exceptions may be too disruptive for the flow of a program or even - render execution impossible. + render execution impossible. For these cases and in order to simplify error handling when dealing with databases, database module authors may choose to @@ -806,7 +806,7 @@ Optional Error Handling Extension Cursor/Connection Attribute .errorhandler Read/write attribute which references an error handler to call - in case an error condition is met. + in case an error condition is met. The handler must be a Python callable taking the following arguments: errorhandler(connection, cursor, errorclass, @@ -836,7 +836,7 @@ Frequently Asked Questions specification. This section covers some of the issues people sometimes have with the specification. - Question: + Question: How can I construct a dictionary out of the tuples returned by .fetchxxx(): @@ -855,7 +855,7 @@ Frequently Asked Questions * Some databases don't support case-sensitive column names or auto-convert them to all lowercase or all uppercase characters. - + * Columns in the result set which are generated by the query (e.g. using SQL functions) don't map to table column names and databases usually generate names for these columns in a @@ -872,9 +872,9 @@ Major Changes from Version 1.0 to Version 2.0 compared to the 1.0 version. Because some of these changes will cause existing DB API 1.0 based scripts to break, the major version number was adjusted to reflect this change. - + These are the most important changes from 1.0 to 2.0: - + * The need for a separate dbi module was dropped and the functionality merged into the module interface itself. @@ -886,10 +886,10 @@ Major Changes from Version 1.0 to Version 2.0 * New constants (apilevel, threadlevel, paramstyle) and methods (executemany, nextset) were added to provide better database bindings. - + * The semantics of .callproc() needed to call stored procedures are now clearly defined. - + * The definition of the .execute() return value changed. Previously, the return value was based on the SQL statement type (which was hard to implement right) -- it is undefined @@ -898,7 +898,7 @@ Major Changes from Version 1.0 to Version 2.0 values, but these are no longer mandated by the specification and should be considered database interface dependent. - + * Class based exceptions were incorporated into the specification. Module implementors are free to extend the exception layout defined in this specification by @@ -916,10 +916,10 @@ Open Issues questions that were left open in the 1.0 version, there are still some remaining issues which should be addressed in future versions: - + * Define a useful return value for .nextset() for the case where a new result set is available. - + * Create a fixed point numeric type for use as loss-less monetary and decimal interchange format. @@ -929,17 +929,17 @@ Footnotes [1] As a guideline the connection constructor parameters should be implemented as keyword parameters for more intuitive use and follow this order of parameters: - + dsn Data source name as string user User name as string (optional) password Password as string (optional) host Hostname (optional) database Database name (optional) - + E.g. a connect could look like this: - + connect(dsn='myhost:MYDB',user='guido',password='234$') - + [2] Module implementors should prefer 'numeric', 'named' or 'pyformat' over the other formats because these offer more clarity and flexibility. @@ -947,41 +947,41 @@ Footnotes [3] If the database does not support the functionality required by the method, the interface should throw an exception in case the method is used. - + The preferred approach is to not implement the method and thus have Python generate an AttributeError in case the method is requested. This allows the programmer to check for database capabilities using the standard hasattr() function. - + For some dynamically configured interfaces it may not be appropriate to require dynamically making the method available. These interfaces should then raise a NotSupportedError to indicate the non-ability to perform the roll back when the method is invoked. - + [4] a database interface may choose to support named cursors by allowing a string argument to the method. This feature is not part of the specification, since it complicates semantics of the .fetchXXX() methods. - + [5] The module will use the __getitem__ method of the parameters object to map either positions (integers) or names (strings) to parameter values. This allows for both sequences and mappings to be used as input. - + The term "bound" refers to the process of binding an input value to a database execution buffer. In practical terms, this means that the input value is directly used as a value in the operation. The client should not be required to "escape" the value so that it can be used -- the value should be equal to the actual database value. - + [6] Note that the interface may implement row fetching using arrays and other optimizations. It is not guaranteed that a call to this method will only move the associated cursor forward by one row. - + [7] The rowcount attribute may be coded in a way that updates its value dynamically. This can be useful for databases that return usable rowcount values only after the first call to diff --git a/doc/release.rst b/doc/release.rst index 00399c7f..3576cdc9 100644 --- a/doc/release.rst +++ b/doc/release.rst @@ -36,7 +36,7 @@ How to make a psycopg2 release - Create a signed tag with the content of the relevant NEWS bit and push it. E.g.:: - $ git tag -a -s 2_7 + $ git tag -a -s 2_7 Psycopg 2.7 released diff --git a/doc/src/advanced.rst b/doc/src/advanced.rst index b2b8afe3..d1683b8b 100644 --- a/doc/src/advanced.rst +++ b/doc/src/advanced.rst @@ -188,7 +188,7 @@ representation into the previously defined `!Point` class: ... return Point(float(m.group(1)), float(m.group(2))) ... else: ... raise InterfaceError("bad point representation: %r" % value) - + In order to create a mapping from a PostgreSQL type (either standard or user-defined), its OID must be known. It can be retrieved either by the second diff --git a/doc/src/connection.rst b/doc/src/connection.rst index 454f80d7..6f13ecd3 100644 --- a/doc/src/connection.rst +++ b/doc/src/connection.rst @@ -22,7 +22,7 @@ The ``connection`` class :ref:`thread-safety` for details. .. method:: cursor(name=None, cursor_factory=None, scrollable=None, withhold=False) - + Return a new `cursor` object using the connection. If *name* is specified, the returned cursor will be a :ref:`server @@ -274,8 +274,8 @@ The ``connection`` class .. __: http://jdbc.postgresql.org/ - Xids returned by `!tpc_recover()` also have extra attributes - `~psycopg2.extensions.Xid.prepared`, `~psycopg2.extensions.Xid.owner`, + Xids returned by `!tpc_recover()` also have extra attributes + `~psycopg2.extensions.Xid.prepared`, `~psycopg2.extensions.Xid.owner`, `~psycopg2.extensions.Xid.database` populated with the values read from the server. @@ -626,7 +626,7 @@ The ``connection`` class pair: Server; Parameters .. method:: get_parameter_status(parameter) - + Look up a current parameter setting of the server. Potential values for ``parameter`` are: ``server_version``, @@ -708,7 +708,7 @@ The ``connection`` class The number is formed by converting the major, minor, and revision numbers into two-decimal-digit numbers and appending them together. For example, version 8.1.5 will be returned as ``80105``. - + .. seealso:: libpq docs for `PQserverVersion()`__ for details. .. __: http://www.postgresql.org/docs/current/static/libpq-status.html#LIBPQ-PQSERVERVERSION @@ -722,7 +722,7 @@ The ``connection`` class .. attribute:: status A read-only integer representing the status of the connection. - Symbolic constants for the values are defined in the module + Symbolic constants for the values are defined in the module `psycopg2.extensions`: see :ref:`connection-status-constants` for the available values. diff --git a/doc/src/cursor.rst b/doc/src/cursor.rst index 417ab6cf..5a6935e6 100644 --- a/doc/src/cursor.rst +++ b/doc/src/cursor.rst @@ -34,10 +34,10 @@ The ``cursor`` class many cursors from the same connection and should use each cursor from a single thread. See :ref:`thread-safety` for details. - - .. attribute:: description - This read-only attribute is a sequence of 7-item sequences. + .. attribute:: description + + This read-only attribute is a sequence of 7-item sequences. Each of these sequences is a named tuple (a regular tuple if :func:`collections.namedtuple` is not available) containing information @@ -65,7 +65,7 @@ The ``cursor`` class This attribute will be `!None` for operations that do not return rows or if the cursor has not had an operation invoked via the |execute*|_ methods yet. - + .. |pg_type| replace:: :sql:`pg_type` .. _pg_type: http://www.postgresql.org/docs/current/static/catalog-pg-type.html .. _PQgetlength: http://www.postgresql.org/docs/current/static/libpq-exec.html#LIBPQ-PQGETLENGTH @@ -78,7 +78,7 @@ The ``cursor`` class regular tuples. .. method:: close() - + Close the cursor now (rather than whenever `del` is executed). The cursor will be unusable from this point forward; an `~psycopg2.InterfaceError` will be raised if any operation is @@ -88,7 +88,7 @@ The ``cursor`` class the method is automatically called at the end of the ``with`` block. - + .. attribute:: closed Read-only boolean attribute: specifies if the cursor is closed @@ -235,7 +235,7 @@ The ``cursor`` class The `mogrify()` method is a Psycopg extension to the |DBAPI|. .. method:: setinputsizes(sizes) - + This method is exposed in compliance with the |DBAPI|. It currently does nothing but it is safe to call it. @@ -281,17 +281,17 @@ The ``cursor`` class >>> cur.execute("SELECT * FROM test WHERE id = %s", (3,)) >>> cur.fetchone() (3, 42, 'bar') - + A `~psycopg2.ProgrammingError` is raised if the previous call to |execute*|_ did not produce any result set or no call was issued yet. .. method:: fetchmany([size=cursor.arraysize]) - + Fetch the next set of rows of a query result, returning a list of tuples. An empty list is returned when no more rows are available. - + The number of rows to fetch per call is specified by the parameter. If it is not given, the cursor's `~cursor.arraysize` determines the number of rows to be fetched. The method should try to fetch as @@ -309,7 +309,7 @@ The ``cursor`` class A `~psycopg2.ProgrammingError` is raised if the previous call to |execute*|_ did not produce any result set or no call was issued yet. - + Note there are performance considerations involved with the size parameter. For optimal performance, it is usually best to use the `~cursor.arraysize` attribute. If the size parameter is used, @@ -344,7 +344,7 @@ The ``cursor`` class `~psycopg2.ProgrammingError` is raised and the cursor position is not changed. - .. note:: + .. note:: According to the |DBAPI|_, the exception raised for a cursor out of bound should have been `!IndexError`. The best option is @@ -364,7 +364,7 @@ The ``cursor`` class .. attribute:: arraysize - + This read/write attribute specifies the number of rows to fetch at a time with `~cursor.fetchmany()`. It defaults to 1 meaning to fetch a single row at a time. @@ -378,20 +378,20 @@ The ``cursor`` class default is 2000. .. versionadded:: 2.4 - + .. extension:: The `itersize` attribute is a Psycopg extension to the |DBAPI|. - .. attribute:: rowcount - + .. attribute:: rowcount + This read-only attribute specifies the number of rows that the last |execute*|_ produced (for :abbr:`DQL (Data Query Language)` statements - like :sql:`SELECT`) or affected (for + like :sql:`SELECT`) or affected (for :abbr:`DML (Data Manipulation Language)` statements like :sql:`UPDATE` or :sql:`INSERT`). - + The attribute is -1 in case no |execute*| has been performed on the cursor or the row count of the last operation if it can't be determined by the interface. @@ -400,7 +400,7 @@ The ``cursor`` class The |DBAPI|_ interface reserves to redefine the latter case to have the object return `!None` instead of -1 in future versions of the specification. - + .. attribute:: rownumber @@ -457,7 +457,7 @@ The ``cursor`` class command: >>> cur.execute("INSERT INTO test (num, data) VALUES (%s, %s)", (42, 'bar')) - >>> cur.statusmessage + >>> cur.statusmessage 'INSERT 0 1' .. extension:: @@ -490,13 +490,13 @@ The ``cursor`` class .. method:: nextset() - + This method is not supported (PostgreSQL does not have multiple data sets) and will raise a `~psycopg2.NotSupportedError` exception. .. method:: setoutputsize(size [, column]) - + This method is exposed in compliance with the |DBAPI|. It currently does nothing but it is safe to call it. diff --git a/doc/src/faq.rst b/doc/src/faq.rst index 0ef4c708..5824d2b4 100644 --- a/doc/src/faq.rst +++ b/doc/src/faq.rst @@ -334,4 +334,3 @@ Psycopg raises *ImportError: cannot import name tz* on import in mod_wsgi / ASP, .. _egg: http://peak.telecommunity.com/DevCenter/PythonEggs .. __: http://stackoverflow.com/questions/2192323/what-is-the-python-egg-cache-python-egg-cache .. __: http://code.google.com/p/modwsgi/wiki/ConfigurationDirectives#WSGIPythonEggs - diff --git a/doc/src/index.rst b/doc/src/index.rst index 30ba8fa1..852bbc2c 100644 --- a/doc/src/index.rst +++ b/doc/src/index.rst @@ -65,4 +65,3 @@ Psycopg 2 is both Unicode and Python 3 friendly. **To Do items in the documentation** .. todolist:: - diff --git a/doc/src/pool.rst b/doc/src/pool.rst index 32bca548..7624dc88 100644 --- a/doc/src/pool.rst +++ b/doc/src/pool.rst @@ -57,8 +57,7 @@ be used. .. autoclass:: PersistentConnectionPool - .. note:: + .. note:: This pool class is mostly designed to interact with Zope and probably not useful in generic applications. - diff --git a/doc/src/tools/lib/dbapi_extension.py b/doc/src/tools/lib/dbapi_extension.py index 7ab656fa..23d91652 100755 --- a/doc/src/tools/lib/dbapi_extension.py +++ b/doc/src/tools/lib/dbapi_extension.py @@ -49,4 +49,3 @@ def setup(app): text=(visit_extension_node, depart_extension_node)) app.add_directive('extension', Extension) - diff --git a/doc/src/tools/lib/sql_role.py b/doc/src/tools/lib/sql_role.py index 8fb8ab84..43347b4a 100644 --- a/doc/src/tools/lib/sql_role.py +++ b/doc/src/tools/lib/sql_role.py @@ -12,10 +12,9 @@ from docutils import nodes, utils from docutils.parsers.rst import roles def sql_role(name, rawtext, text, lineno, inliner, options={}, content=[]): - text = utils.unescape(text) + text = utils.unescape(text) options['classes'] = ['sql'] return [nodes.literal(rawtext, text, **options)], [] - + def setup(app): roles.register_local_role('sql', sql_role) - diff --git a/doc/src/tools/lib/ticket_role.py b/doc/src/tools/lib/ticket_role.py index d8ded227..0ee3d636 100644 --- a/doc/src/tools/lib/ticket_role.py +++ b/doc/src/tools/lib/ticket_role.py @@ -56,4 +56,3 @@ def setup(app): app.add_config_value('ticket_remap_offset', None, 'env') app.add_role('ticket', ticket_role) app.add_role('tickets', ticket_role) - diff --git a/doc/src/tools/stitch_text.py b/doc/src/tools/stitch_text.py index c9ed99aa..b36727b5 100755 --- a/doc/src/tools/stitch_text.py +++ b/doc/src/tools/stitch_text.py @@ -57,7 +57,6 @@ def emit(basename, txt_dir): # some space between sections sys.stdout.write("\n\n") - + if __name__ == '__main__': sys.exit(main()) - diff --git a/doc/src/tz.rst b/doc/src/tz.rst index 6e2e9811..06333f86 100644 --- a/doc/src/tz.rst +++ b/doc/src/tz.rst @@ -8,9 +8,8 @@ This module holds two different tzinfo implementations that can be used as the `tzinfo` argument to `~datetime.datetime` constructors, directly passed to Psycopg functions or used to set the `cursor.tzinfo_factory` attribute in -cursors. +cursors. .. autoclass:: psycopg2.tz.FixedOffsetTimezone .. autoclass:: psycopg2.tz.LocalTimezone - diff --git a/doc/src/usage.rst b/doc/src/usage.rst index 5dcab8fe..e52bb9fa 100644 --- a/doc/src/usage.rst +++ b/doc/src/usage.rst @@ -1017,4 +1017,3 @@ For further details see the documentation for the above methods. .. __: http://www.opengroup.org/bookstore/catalog/c193.htm .. __: http://jdbc.postgresql.org/ - diff --git a/examples/binary.py b/examples/binary.py index 3ba1ff89..665d962f 100644 --- a/examples/binary.py +++ b/examples/binary.py @@ -62,7 +62,7 @@ for row in curs.fetchall(): open(new_name, 'wb').write(row[2]) print "done" print " python type of image data is", type(row[2]) - + # extract exactly the same data but using a binary cursor print "Extracting the images using a binary cursor:" @@ -78,7 +78,7 @@ for row in curs.fetchall(): open(new_name, 'wb').write(row[0]) print "done" print " python type of image data is", type(row[0]) - + # this rollback is required because we can't drop a table with a binary cursor # declared and still open conn.rollback() diff --git a/examples/copy_from.py b/examples/copy_from.py index e258a077..57986dbf 100644 --- a/examples/copy_from.py +++ b/examples/copy_from.py @@ -1,4 +1,4 @@ -# copy_from.py -- example about copy_from +# copy_from.py -- example about copy_from # # Copyright (C) 2002 Tom Jenkins # Copyright (C) 2005 Federico Di Gregorio @@ -172,6 +172,3 @@ conn.rollback() curs.execute("DROP TABLE test_copy") os.unlink('copy_from.txt') conn.commit() - - - diff --git a/examples/copy_to.py b/examples/copy_to.py index 225155c1..34c90715 100644 --- a/examples/copy_to.py +++ b/examples/copy_to.py @@ -1,4 +1,4 @@ -# copy_to.py -- example about copy_to +# copy_to.py -- example about copy_to # # Copyright (C) 2002 Tom Jenkins # Copyright (C) 2005 Federico Di Gregorio diff --git a/examples/cursor.py b/examples/cursor.py index 465967c3..58c8cf67 100644 --- a/examples/cursor.py +++ b/examples/cursor.py @@ -49,7 +49,7 @@ class Cursor(psycopg2.extensions.cursor): if d is None: raise NoDataError("no more data") return d - + curs = conn.cursor(cursor_factory=Cursor) curs.execute("SELECT 1 AS foo") print("Result of fetchone():", curs.fetchone()) diff --git a/examples/dialtone.py b/examples/dialtone.py index a89021ce..8654ffde 100644 --- a/examples/dialtone.py +++ b/examples/dialtone.py @@ -6,14 +6,14 @@ Mapping arbitrary objects to a PostgreSQL database with psycopg2 - Problem You need to store arbitrary objects in a PostgreSQL database without being -intrusive for your classes (don't want inheritance from an 'Item' or +intrusive for your classes (don't want inheritance from an 'Item' or 'Persistent' object). - Solution """ from datetime import datetime - + import psycopg2 from psycopg2.extensions import adapt, register_adapter @@ -24,7 +24,7 @@ except: seq.sort() return seq -# Here is the adapter for every object that we may ever need to +# Here is the adapter for every object that we may ever need to # insert in the database. It receives the original object and does # its job on that instance @@ -33,7 +33,7 @@ class ObjectMapper(object): self.orig = orig self.tmp = {} self.items, self.fields = self._gatherState() - + def _gatherState(self): adaptee_name = self.orig.__class__.__name__ fields = sorted([(field, getattr(self.orig, field)) @@ -42,19 +42,19 @@ class ObjectMapper(object): for item, value in fields: items.append(item) return items, fields - + def getTableName(self): return self.orig.__class__.__name__ - + def getMappedValues(self): tmp = [] for i in self.items: tmp.append("%%(%s)s"%i) return ", ".join(tmp) - + def getValuesDict(self): return dict(self.fields) - + def getFields(self): return self.items @@ -66,14 +66,14 @@ class ObjectMapper(object): return qry, self.getValuesDict() # Here are the objects -class Album(object): - id = 0 +class Album(object): + id = 0 def __init__(self): self.creation_time = datetime.now() self.album_id = self.id Album.id = Album.id + 1 self.binary_data = buffer('12312312312121') - + class Order(object): id = 0 def __init__(self): @@ -84,7 +84,7 @@ class Order(object): register_adapter(Album, ObjectMapper) register_adapter(Order, ObjectMapper) - + # Describe what is needed to save on each object # This is actually just configuration, you can use xml with a parser if you # like to have plenty of wasted CPU cycles ;P. @@ -92,7 +92,7 @@ register_adapter(Order, ObjectMapper) persistent_fields = {'Album': ['album_id', 'creation_time', 'binary_data'], 'Order': ['order_id', 'items', 'price'] } - + print adapt(Album()).generateInsert() print adapt(Album()).generateInsert() print adapt(Album()).generateInsert() @@ -103,42 +103,42 @@ print adapt(Order()).generateInsert() """ - Discussion -Psycopg 2 has a great new feature: adaptation. The big thing about -adaptation is that it enables the programmer to glue most of the +Psycopg 2 has a great new feature: adaptation. The big thing about +adaptation is that it enables the programmer to glue most of the code out there without many difficulties. -This recipe tries to focus attention on a way to generate SQL queries to -insert completely new objects inside a database. As you can see objects do -not know anything about the code that is handling them. We specify all the +This recipe tries to focus attention on a way to generate SQL queries to +insert completely new objects inside a database. As you can see objects do +not know anything about the code that is handling them. We specify all the fields that we need for each object through the persistent_fields dict. The most important lines of this recipe are: register_adapter(Album, ObjectMapper) register_adapter(Order, ObjectMapper) -In these lines we notify the system that when we call adapt with an Album instance -as an argument we want it to istantiate ObjectMapper passing the Album instance +In these lines we notify the system that when we call adapt with an Album instance +as an argument we want it to istantiate ObjectMapper passing the Album instance as argument (self.orig in the ObjectMapper class). The output is something like this (for each call to generateInsert): - -('INSERT INTO Album (album_id, binary_data, creation_time) VALUES - (%(album_id)s, %(binary_data)s, %(creation_time)s)', - - {'binary_data': , - 'creation_time': datetime.datetime(2004, 9, 10, 20, 48, 29, 633728), + +('INSERT INTO Album (album_id, binary_data, creation_time) VALUES + (%(album_id)s, %(binary_data)s, %(creation_time)s)', + + {'binary_data': , + 'creation_time': datetime.datetime(2004, 9, 10, 20, 48, 29, 633728), 'album_id': 1} ) -This is a tuple of {SQL_QUERY, FILLING_DICT}, and all the quoting/converting -stuff (from python's datetime to postgres s and from python's buffer to -postgres' blob) is handled with the same adaptation process hunder the hood +This is a tuple of {SQL_QUERY, FILLING_DICT}, and all the quoting/converting +stuff (from python's datetime to postgres s and from python's buffer to +postgres' blob) is handled with the same adaptation process hunder the hood by psycopg2. -At last, just notice that ObjectMapper is working for both Album and Order -instances without any glitches at all, and both classes could have easily been -coming from closed source libraries or C coded ones (which are not easily -modified), whereas a common pattern in todays ORMs or OODBs is to provide -a basic 'Persistent' object that already knows how to store itself in the +At last, just notice that ObjectMapper is working for both Album and Order +instances without any glitches at all, and both classes could have easily been +coming from closed source libraries or C coded ones (which are not easily +modified), whereas a common pattern in todays ORMs or OODBs is to provide +a basic 'Persistent' object that already knows how to store itself in the database. """ diff --git a/examples/dict.py b/examples/dict.py index ede014a5..1503722c 100644 --- a/examples/dict.py +++ b/examples/dict.py @@ -29,7 +29,7 @@ print "Opening connection using dsn:", DSN conn = psycopg2.connect(DSN) print "Encoding for this connection is", conn.encoding - + curs = conn.cursor(cursor_factory=psycopg2.extras.DictCursor) curs.execute("SELECT 1 AS foo, 'cip' AS bar, date(now()) as zot") print "Cursor's row factory is", curs.row_factory diff --git a/examples/encoding.py b/examples/encoding.py index c702b054..77fd871b 100644 --- a/examples/encoding.py +++ b/examples/encoding.py @@ -38,7 +38,7 @@ encs.sort() for a, b in encs: print " ", a, "<->", b -print "Using STRING typecaster" +print "Using STRING typecaster" print "Setting backend encoding to LATIN1 and executing queries:" conn.set_client_encoding('LATIN1') curs = conn.cursor() diff --git a/examples/myfirstrecipe.py b/examples/myfirstrecipe.py index c6680c31..e0df879f 100644 --- a/examples/myfirstrecipe.py +++ b/examples/myfirstrecipe.py @@ -34,7 +34,7 @@ a not-yet well defined protocol that we can call ISQLQuote: def getbinary(self): "Returns a binary quoted string representing the bound variable." - + def getbuffer(self): "Returns the wrapped object itself." @@ -86,10 +86,10 @@ class AsIs(object): self.__obj = obj def getquoted(self): return self.__obj - + class SQL_IN(object): """Adapt a tuple to an SQL quotable object.""" - + def __init__(self, seq): self._seq = seq @@ -103,10 +103,10 @@ class SQL_IN(object): qobjs = [str(psycoadapt(o).getquoted()) for o in self._seq] return '(' + ', '.join(qobjs) + ')' - + __str__ = getquoted - + # add our new adapter class to psycopg list of adapters register_adapter(tuple, SQL_IN) register_adapter(float, AsIs) @@ -117,7 +117,7 @@ register_adapter(int, AsIs) # conn = psycopg.connect("...") # curs = conn.cursor() # curs.execute("SELECT ...", (("this", "is", "the", "tuple"),)) -# +# # but we have no connection to a database right now, so we just check # the SQL_IN class by calling psycopg's adapt() directly: diff --git a/examples/threads.py b/examples/threads.py index d24b0a5d..e0be7e0c 100644 --- a/examples/threads.py +++ b/examples/threads.py @@ -44,7 +44,7 @@ if len(sys.argv) > 1: DSN = sys.argv[1] if len(sys.argv) > 2: MODE = int(sys.argv[2]) - + print "Opening connection using dsn:", DSN conn = psycopg2.connect(DSN) curs = conn.cursor() @@ -70,7 +70,7 @@ def insert_func(conn_or_pool, rows): conn = conn_or_pool else: conn = conn_or_pool.getconn() - + for i in range(rows): if divmod(i, COMMIT_STEP)[1] == 0: conn.commit() @@ -91,14 +91,14 @@ def insert_func(conn_or_pool, rows): ## a nice select function that prints the current number of rows in the ## database (and transfer them, putting some pressure on the network) - + def select_func(conn_or_pool, z): name = threading.currentThread().getName() if MODE == 0: conn = conn_or_pool conn.set_isolation_level(ISOLATION_LEVEL_AUTOCOMMIT) - + for i in range(SELECT_SIZE): if divmod(i, SELECT_STEP)[1] == 0: try: @@ -125,7 +125,7 @@ else: m = len(INSERT_THREADS) + len(SELECT_THREADS) n = m/2 conn_insert = conn_select = ThreadedConnectionPool(n, m, DSN) - + ## create the threads threads = [] diff --git a/examples/typecast.py b/examples/typecast.py index 0df883e7..169f3aca 100644 --- a/examples/typecast.py +++ b/examples/typecast.py @@ -63,5 +63,3 @@ print "Some text from cursor with typecaster:", curs.fetchone()[0] curs = conn.cursor() curs.execute("SELECT 'some text.'::text AS foo") print "Some text from connection with typecaster again:", curs.fetchone()[0] - - diff --git a/examples/tz.py b/examples/tz.py index 9a484f1e..c3dd8e43 100644 --- a/examples/tz.py +++ b/examples/tz.py @@ -60,10 +60,10 @@ print "Time zone:", d.tzinfo.tzname(d), "offset:", d.tzinfo.utcoffset(d) curs.execute("SELECT * FROM test_tz") for d in curs: u = d[0].utcoffset() or ZERO - print "UTC time: ", d[0] - u + print "UTC time: ", d[0] - u print "Local time:", d[0] print "Time zone:", d[0].tzinfo.tzname(d[0]), d[0].tzinfo.utcoffset(d[0]) - + curs.execute("DROP TABLE test_tz") conn.commit() diff --git a/examples/usercast.py b/examples/usercast.py index 87c2eea4..26a5c16e 100644 --- a/examples/usercast.py +++ b/examples/usercast.py @@ -58,7 +58,7 @@ class Rect(object): and eventually as a type-caster for the data extracted from the database (that's why __init__ takes the curs argument.) """ - + def __init__(self, s=None, curs=None): """Init the rectangle from the optional string s.""" self.x = self.y = self.width = self.height = 0.0 @@ -68,7 +68,7 @@ class Rect(object): """This is a terrible hack, just ignore proto and return self.""" if proto == psycopg2.extensions.ISQLQuote: return self - + def from_points(self, x0, y0, x1, y1): """Init the rectangle from points.""" if x0 > x1: (x0, x1) = (x1, x0) @@ -94,7 +94,7 @@ class Rect(object): s = "X: %d\tY: %d\tWidth: %d\tHeight: %d" % ( self.x, self.y, self.width, self.height) return s - + # here we select from the empty table, just to grab the description curs.execute("SELECT b FROM test_cast WHERE 0=1") boxoid = curs.description[0][1] diff --git a/psycopg/bytes_format.c b/psycopg/bytes_format.c index 68662075..12899142 100644 --- a/psycopg/bytes_format.c +++ b/psycopg/bytes_format.c @@ -295,4 +295,3 @@ Bytes_Format(PyObject *format, PyObject *args) } return NULL; } - diff --git a/psycopg/notify_type.c b/psycopg/notify_type.c index e2589a6a..aab9a43b 100644 --- a/psycopg/notify_type.c +++ b/psycopg/notify_type.c @@ -295,5 +295,3 @@ PyTypeObject notifyType = { 0, /*tp_alloc*/ notify_new, /*tp_new*/ }; - - diff --git a/psycopg/typecast_binary.c b/psycopg/typecast_binary.c index 8c380239..e4839dab 100644 --- a/psycopg/typecast_binary.c +++ b/psycopg/typecast_binary.c @@ -312,4 +312,3 @@ psycopg_parse_escape(const char *bufin, Py_ssize_t sizein, Py_ssize_t *sizeout) exit: return ret; } - diff --git a/psycopg/typecast_builtins.c b/psycopg/typecast_builtins.c index 1c4a8ca0..446dd14b 100644 --- a/psycopg/typecast_builtins.c +++ b/psycopg/typecast_builtins.c @@ -69,4 +69,3 @@ static typecastObject_initlist typecast_builtins[] = { {"MACADDRARRAY", typecast_MACADDRARRAY_types, typecast_STRINGARRAY_cast, "STRING"}, {NULL, NULL, NULL, NULL} }; - diff --git a/psycopg/typecast_mxdatetime.c b/psycopg/typecast_mxdatetime.c index 12c734a6..637e4384 100644 --- a/psycopg/typecast_mxdatetime.c +++ b/psycopg/typecast_mxdatetime.c @@ -250,4 +250,3 @@ typecast_MXINTERVAL_cast(const char *str, Py_ssize_t len, PyObject *curs) #define typecast_DATETIME_cast typecast_MXDATE_cast #define typecast_DATETIMETZ_cast typecast_MXDATE_cast #endif - diff --git a/psycopg2.sln b/psycopg2.sln index 829468d2..68327dd8 100644 --- a/psycopg2.sln +++ b/psycopg2.sln @@ -1,5 +1,4 @@ - -Microsoft Visual Studio Solution File, Format Version 10.00 +Microsoft Visual Studio Solution File, Format Version 10.00 # Visual Studio 2008 Project("{2857B73E-F847-4B02-9238-064979017E93}") = "psycopg2", "psycopg2.cproj", "{CFD80D18-3EE5-49ED-992A-E6D433BC7641}" EndProject @@ -26,7 +25,7 @@ Global $2.DirectoryNamespaceAssociation = None $2.ResourceNamePolicy = FileName $0.StandardHeader = $3 - $3.Text = + $3.Text = $3.IncludeInNewFiles = False $0.TextStylePolicy = $4 $4.FileWidth = 72 diff --git a/sandbox/array.py b/sandbox/array.py index 2327a273..d7971561 100644 --- a/sandbox/array.py +++ b/sandbox/array.py @@ -28,4 +28,3 @@ curs = conn.cursor() curs.execute("SELECT %s", ([1,2,None],)) print curs.fetchone() - diff --git a/sandbox/async.py b/sandbox/async.py index 3766fbb9..e9fe2811 100644 --- a/sandbox/async.py +++ b/sandbox/async.py @@ -17,7 +17,7 @@ def sleep(curs): while not curs.isready(): print "." time.sleep(.1) - + #curs.execute(""" # DECLARE zz INSENSITIVE SCROLL CURSOR WITH HOLD FOR # SELECT now(); @@ -33,4 +33,3 @@ print curs.fetchall() curs.execute("SELECT now() AS bar") sleep(curs) - diff --git a/sandbox/gtk.py b/sandbox/gtk.py index f96e1a32..b00cd29b 100644 --- a/sandbox/gtk.py +++ b/sandbox/gtk.py @@ -16,4 +16,3 @@ sql() import gtk print "AFTER" sql() - diff --git a/sandbox/iter.py b/sandbox/iter.py index 74c8aba8..bf245ba9 100644 --- a/sandbox/iter.py +++ b/sandbox/iter.py @@ -7,8 +7,7 @@ curs = conn.cursor(cursor_factory=psycopg2.extras.DictCursor) curs.execute("SELECT '2005-2-12'::date AS foo, 'boo!' as bar") for x in curs.fetchall(): print type(x), x[0], x[1], x['foo'], x['bar'] - + curs.execute("SELECT '2005-2-12'::date AS foo, 'boo!' as bar") for x in curs: print type(x), x[0], x[1], x['foo'], x['bar'] - diff --git a/sandbox/leak.test.py b/sandbox/leak.test.py index 040f31a2..c0bef371 100644 --- a/sandbox/leak.test.py +++ b/sandbox/leak.test.py @@ -14,7 +14,7 @@ two functions: # leak() will cause increasingly more RAM to be used by the script. $ python leak -# noleak() does not have the RAM usage problem. The only difference +# noleak() does not have the RAM usage problem. The only difference # between it and leak() is that 'rows' is created once, before the loop. $ python noleak @@ -80,4 +80,3 @@ except IndexError: # Run leak() or noleak(), whichever was indicated on the command line run_function() - diff --git a/sandbox/misc_dbapi_test.py b/sandbox/misc_dbapi_test.py index 64637d98..137ddb04 100644 --- a/sandbox/misc_dbapi_test.py +++ b/sandbox/misc_dbapi_test.py @@ -28,8 +28,8 @@ import psycopg2 as dbapi conn = dbapi.connect(database='test') - - + + cursor = conn.cursor() cursor.execute(""" @@ -41,5 +41,3 @@ cursor.execute(""" for row in cursor: print row - - diff --git a/sandbox/named.py b/sandbox/named.py index 6a107604..94d89811 100644 --- a/sandbox/named.py +++ b/sandbox/named.py @@ -5,7 +5,7 @@ class Portal(psycopg2.extensions.cursor): def __init__(self, name, curs): psycopg2.extensions.cursor.__init__( self, curs.connection, '"'+name+'"') - + CURSOR = psycopg2.extensions.new_type((1790,), "CURSOR", Portal) psycopg2.extensions.register_type(CURSOR) diff --git a/sandbox/pbool.py b/sandbox/pbool.py index 817e3ea0..d26f8b46 100644 --- a/sandbox/pbool.py +++ b/sandbox/pbool.py @@ -10,4 +10,3 @@ class B(object): return 'It is True' else: return 'It is False' - diff --git a/sandbox/test.py b/sandbox/test.py index e3d31a75..a2a2a6b6 100644 --- a/sandbox/test.py +++ b/sandbox/test.py @@ -31,7 +31,7 @@ def sleep(curs): while not curs.isready(): print "." time.sleep(.1) - + #curs.execute(""" # DECLARE zz INSENSITIVE SCROLL CURSOR WITH HOLD FOR # SELECT now(); @@ -47,4 +47,3 @@ def sleep(curs): #curs.execute("SELECT now() AS bar"); #sleep(curs) - diff --git a/sandbox/test814.py b/sandbox/test814.py index 53e7e7f8..1e20ba0d 100644 --- a/sandbox/test814.py +++ b/sandbox/test814.py @@ -6,4 +6,3 @@ curs = conn.cursor() curs.execute("SELECT true AS foo WHERE 'a' in %s", (("aa", "bb"),)) print curs.fetchall() print curs.query - diff --git a/sandbox/test_copy2.py b/sandbox/test_copy2.py index bdd6e9e6..088c17b3 100644 --- a/sandbox/test_copy2.py +++ b/sandbox/test_copy2.py @@ -40,4 +40,3 @@ dbconn.commit() cursor.close() dbconn.close() - diff --git a/sandbox/textfloat.py b/sandbox/textfloat.py index 22d65b8b..5383bb38 100644 --- a/sandbox/textfloat.py +++ b/sandbox/textfloat.py @@ -6,4 +6,3 @@ c = o.cursor() c.execute("SELECT 1.23::float AS foo") x = c.fetchone()[0] print x, type(x) - diff --git a/sandbox/ticket58.py b/sandbox/ticket58.py index 95520c1e..98b5452b 100644 --- a/sandbox/ticket58.py +++ b/sandbox/ticket58.py @@ -71,5 +71,3 @@ done = 1 cur.close() conn.close() - - diff --git a/sandbox/valgrind-python.supp b/sandbox/valgrind-python.supp index 0938184f..4364ae75 100644 --- a/sandbox/valgrind-python.supp +++ b/sandbox/valgrind-python.supp @@ -335,7 +335,7 @@ { Debian unstable with libc-i686 suppressions - Memcheck:Cond + Memcheck:Cond obj:/lib/ld-2.3.5.so obj:/lib/ld-2.3.5.so obj:/lib/tls/i686/cmov/libc-2.3.5.so @@ -348,10 +348,10 @@ fun:_PyImport_GetDynLoadFunc fun:_PyImport_LoadDynamicModule } - + { Debian unstable with libc-i686 suppressions - Memcheck:Cond + Memcheck:Cond obj:/lib/ld-2.3.5.so obj:/lib/ld-2.3.5.so obj:/lib/ld-2.3.5.so @@ -365,7 +365,7 @@ fun:_PyImport_GetDynLoadFunc fun:_PyImport_LoadDynamicModule } - + { Debian unstable with libc-i686 suppressions Memcheck:Addr4 @@ -471,7 +471,7 @@ { Debian unstable with libc-i686 suppressions Memcheck:Cond - obj:/lib/ld-2.3.5.so + obj:/lib/ld-2.3.5.so obj:/lib/tls/i686/cmov/libc-2.3.5.so obj:/lib/ld-2.3.5.so fun:_dl_open diff --git a/scripts/maketypes.sh b/scripts/maketypes.sh index ab133c17..4165c1fb 100644 --- a/scripts/maketypes.sh +++ b/scripts/maketypes.sh @@ -28,7 +28,7 @@ PGMINOR="`echo $PGVERSION | cut -d. -f2`" echo checking for postgresql major: $PGMAJOR echo checking for postgresql minor: $PGMINOR - + echo -n generating pgtypes.h ... awk '/#define .+OID/ {print "#define " $2 " " $3}' "$PGTYPE" \ > $SRCDIR/pgtypes.h @@ -37,5 +37,3 @@ echo -n generating typecast_builtins.c ... awk '/#define .+OID/ {print $2 " " $3}' "$PGTYPE" | \ python $SCRIPTSDIR/buildtypes.py >$SRCDIR/typecast_builtins.c echo " done" - - diff --git a/tests/dbapi20.py b/tests/dbapi20.py index ff98ddc6..4facf8e8 100644 --- a/tests/dbapi20.py +++ b/tests/dbapi20.py @@ -1,6 +1,6 @@ #!/usr/bin/env python -''' Python DB API 2.0 driver compliance unit test suite. - +''' Python DB API 2.0 driver compliance unit test suite. + This software is Public Domain and may be used without restrictions. "Now we have booze and barflies entering the discussion, plus rumours of @@ -79,8 +79,8 @@ def str2bytes(sval): class DatabaseAPI20Test(unittest.TestCase): ''' Test a database self.driver for DB API 2.0 compatibility. This implementation tests Gadfly, but the TestCase - is structured so that other self.drivers can subclass this - test case to ensure compiliance with the DB-API. It is + is structured so that other self.drivers can subclass this + test case to ensure compiliance with the DB-API. It is expected that this TestCase may be expanded in the future if ambiguities or edge conditions are discovered. @@ -90,9 +90,9 @@ class DatabaseAPI20Test(unittest.TestCase): self.driver, connect_args and connect_kw_args. Class specification should be as follows: - import dbapi20 + import dbapi20 class mytest(dbapi20.DatabaseAPI20Test): - [...] + [...] Don't 'import DatabaseAPI20Test from dbapi20', or you will confuse the unit tester - just 'import dbapi20'. @@ -111,7 +111,7 @@ class DatabaseAPI20Test(unittest.TestCase): xddl2 = 'drop table %sbarflys' % table_prefix lowerfunc = 'lower' # Name of stored procedure to convert string->lowercase - + # Some drivers may need to override these helpers, for example adding # a 'commit' after the execute. def executeDDL1(self,cursor): @@ -135,10 +135,10 @@ class DatabaseAPI20Test(unittest.TestCase): try: cur = con.cursor() for ddl in (self.xddl1,self.xddl2): - try: + try: cur.execute(ddl) con.commit() - except self.driver.Error: + except self.driver.Error: # Assume table didn't exist. Other tests will check if # execute is busted. pass @@ -255,7 +255,7 @@ class DatabaseAPI20Test(unittest.TestCase): con.rollback() except self.driver.NotSupportedError: pass - + def test_cursor(self): con = self._connect() try: @@ -411,7 +411,7 @@ class DatabaseAPI20Test(unittest.TestCase): ) elif self.driver.paramstyle == 'named': cur.execute( - 'insert into %sbooze values (:beer)' % self.table_prefix, + 'insert into %sbooze values (:beer)' % self.table_prefix, {'beer':"Cooper's"} ) elif self.driver.paramstyle == 'format': @@ -551,7 +551,7 @@ class DatabaseAPI20Test(unittest.TestCase): tests. ''' populate = [ - "insert into %sbooze values ('%s')" % (self.table_prefix,s) + "insert into %sbooze values ('%s')" % (self.table_prefix,s) for s in self.samples ] return populate @@ -612,7 +612,7 @@ class DatabaseAPI20Test(unittest.TestCase): self.assertEqual(len(rows),6) rows = [r[0] for r in rows] rows.sort() - + # Make sure we get the right data back out for i in range(0,6): self.assertEqual(rows[i],self.samples[i], @@ -683,10 +683,10 @@ class DatabaseAPI20Test(unittest.TestCase): 'cursor.fetchall should return an empty list if ' 'a select query returns no rows' ) - + finally: con.close() - + def test_mixedfetch(self): con = self._connect() try: @@ -722,7 +722,7 @@ class DatabaseAPI20Test(unittest.TestCase): def help_nextset_setUp(self,cur): ''' Should create a procedure called deleteme - that returns two result sets, first the + that returns two result sets, first the number of rows in booze then "name from booze" ''' raise NotImplementedError('Helper not implemented') @@ -869,4 +869,3 @@ class DatabaseAPI20Test(unittest.TestCase): self.failUnless(hasattr(self.driver,'ROWID'), 'module.ROWID must be defined.' ) -