docs: fix a few simple typos

There are small typos in:
- doc/src/faq.rst
- psycopg/cursor_type.c
- psycopg/xid_type.c
- scripts/make_errorcodes.py
- scripts/make_errors.py
- tests/dbapi20.py
- tests/test_connection.py

Fixes:
- Should read `publicly` rather than `publically`.
- Should read `unicode` rather than `uniconde`.
- Should read `supposed` rather than `suppsed`.
- Should read `something` rather than `somenthing`.
- Should read `portable` rather than `portible`.
- Should read `exhausted` rather than `exausted`.
- Should read `compliance` rather than `compiliance`.

Closes #1275
This commit is contained in:
Tim Gates 2021-05-18 05:47:41 +10:00 committed by Daniele Varrazzo
parent 9d8c7a7036
commit ba637a5e0c
7 changed files with 8 additions and 8 deletions

View File

@ -180,7 +180,7 @@ Psycopg automatically converts PostgreSQL :sql:`json` data into Python objects.
Psycopg converts :sql:`json` values into Python objects but :sql:`jsonb` values are returned as strings. Can :sql:`jsonb` be converted automatically? Psycopg converts :sql:`json` values into Python objects but :sql:`jsonb` values are returned as strings. Can :sql:`jsonb` be converted automatically?
Automatic conversion of :sql:`jsonb` values is supported from Psycopg Automatic conversion of :sql:`jsonb` values is supported from Psycopg
release 2.5.4. For previous versions you can register the :sql:`json` release 2.5.4. For previous versions you can register the :sql:`json`
typecaster on the :sql:`jsonb` oids (which are known and not suppsed to typecaster on the :sql:`jsonb` oids (which are known and not supposed to
change in future PostgreSQL versions):: change in future PostgreSQL versions)::
psycopg2.extras.register_json(oid=3802, array_oid=3807, globally=True) psycopg2.extras.register_json(oid=3802, array_oid=3807, globally=True)

View File

@ -768,7 +768,7 @@ curs_fetchone(cursorObject *self, PyObject *dummy)
Dprintf("curs_fetchone: rowcount = %ld", self->rowcount); Dprintf("curs_fetchone: rowcount = %ld", self->rowcount);
if (self->row >= self->rowcount) { if (self->row >= self->rowcount) {
/* we exausted available data: return None */ /* we exhausted available data: return None */
Py_RETURN_NONE; Py_RETURN_NONE;
} }

View File

@ -598,7 +598,7 @@ xid_recover(PyObject *conn)
/* curs = conn.cursor() /* curs = conn.cursor()
* (sort of. Use the real cursor in case the connection returns * (sort of. Use the real cursor in case the connection returns
* somenthing non-dbapi -- see ticket #114) */ * something non-dbapi -- see ticket #114) */
if (!(curs = PyObject_CallFunctionObjArgs( if (!(curs = PyObject_CallFunctionObjArgs(
(PyObject *)&cursorType, conn, NULL))) { goto exit; } (PyObject *)&cursorType, conn, NULL))) { goto exit; }

View File

@ -77,7 +77,7 @@ def parse_errors_txt(url):
m = re.match(r"(.....)\s+(?:E|W|S)\s+ERRCODE_(\S+)(?:\s+(\S+))?$", line) m = re.match(r"(.....)\s+(?:E|W|S)\s+ERRCODE_(\S+)(?:\s+(\S+))?$", line)
if m: if m:
errcode, macro, spec = m.groups() errcode, macro, spec = m.groups()
# skip errcodes without specs as they are not publically visible # skip errcodes without specs as they are not publicly visible
if not spec: if not spec:
continue continue
errlabel = spec.upper() errlabel = spec.upper()

View File

@ -61,7 +61,7 @@ def parse_errors_txt(url):
m = re.match(r"(.....)\s+(?:E|W|S)\s+ERRCODE_(\S+)(?:\s+(\S+))?$", line) m = re.match(r"(.....)\s+(?:E|W|S)\s+ERRCODE_(\S+)(?:\s+(\S+))?$", line)
if m: if m:
errcode, macro, spec = m.groups() errcode, macro, spec = m.groups()
# skip errcodes without specs as they are not publically visible # skip errcodes without specs as they are not publicly visible
if not spec: if not spec:
continue continue
errlabel = spec.upper() errlabel = spec.upper()

View File

@ -64,7 +64,7 @@ import sys
# - self.populate is now self._populate(), so if a driver stub # - self.populate is now self._populate(), so if a driver stub
# overrides self.ddl1 this change propagates # overrides self.ddl1 this change propagates
# - VARCHAR columns now have a width, which will hopefully make the # - VARCHAR columns now have a width, which will hopefully make the
# DDL even more portible (this will be reversed if it causes more problems) # DDL even more portable (this will be reversed if it causes more problems)
# - cursor.rowcount being checked after various execute and fetchXXX methods # - cursor.rowcount being checked after various execute and fetchXXX methods
# - Check for fetchall and fetchmany returning empty lists after results # - Check for fetchall and fetchmany returning empty lists after results
# are exhausted (already checking for empty lists if select retrieved # are exhausted (already checking for empty lists if select retrieved
@ -76,7 +76,7 @@ class DatabaseAPI20Test(unittest.TestCase):
''' Test a database self.driver for DB API 2.0 compatibility. ''' Test a database self.driver for DB API 2.0 compatibility.
This implementation tests Gadfly, but the TestCase This implementation tests Gadfly, but the TestCase
is structured so that other self.drivers can subclass this is structured so that other self.drivers can subclass this
test case to ensure compiliance with the DB-API. It is test case to ensure compliance with the DB-API. It is
expected that this TestCase may be expanded in the future expected that this TestCase may be expanded in the future
if ambiguities or edge conditions are discovered. if ambiguities or edge conditions are discovered.

View File

@ -1242,7 +1242,7 @@ class ConnectionTwoPhaseTests(ConnectingTestCase):
def test_xid_unicode_unparsed(self): def test_xid_unicode_unparsed(self):
# We don't expect people shooting snowmen as transaction ids, # We don't expect people shooting snowmen as transaction ids,
# so if something explodes in an encode error I don't mind. # so if something explodes in an encode error I don't mind.
# Let's just check uniconde is accepted as type. # Let's just check unicode is accepted as type.
cnn = self.connect() cnn = self.connect()
cnn.set_client_encoding('utf8') cnn.set_client_encoding('utf8')
cnn.tpc_begin(u"transaction-id") cnn.tpc_begin(u"transaction-id")