25 KiB
Class Reference
infi.clickhouse_orm.database
Database
Database instances connect to a specific ClickHouse database for running queries, inserting data and other operations.
Database(db_name, db_url="http://localhost:8123/", username=None, password=None, readonly=False, autocreate=True, timeout=60, verify_ssl_cert=True, log_statements=False)
Initializes a database instance. Unless it's readonly, the database will be created on the ClickHouse server if it does not already exist.
db_name
: name of the database to connect to.db_url
: URL of the ClickHouse server.username
: optional connection credentials.password
: optional connection credentials.readonly
: use a read-only connection.autocreate
: automatically create the database if it does not exist (unless in readonly mode).timeout
: the connection timeout in seconds.verify_ssl_cert
: whether to verify the server's certificate when connecting via HTTPS.log_statements
: when True, all database statements are logged.
add_setting(name, value)
Adds a database setting that will be sent with every request.
For example, db.add_setting("max_execution_time", 10)
will
limit query execution time to 10 seconds.
The name must be string, and the value is converted to string in case
it isn't. To remove a setting, pass None
as the value.
count(model_class, conditions=None)
Counts the number of records in the model's table.
model_class
: the model to count.conditions
: optional SQL conditions (contents of the WHERE clause).
create_database()
Creates the database on the ClickHouse server if it does not already exist.
create_table(model_class)
Creates a table for the given model class, if it does not exist already.
does_table_exist(model_class)
Checks whether a table for the given model class already exists. Note that this only checks for existence of a table with the expected name.
drop_database()
Deletes the database on the ClickHouse server.
drop_table(model_class)
Drops the database table of the given model class, if it exists.
insert(model_instances, batch_size=1000)
Insert records into the database.
model_instances
: any iterable containing instances of a single model class.batch_size
: number of records to send per chunk (use a lower number if your records are very large).
migrate(migrations_package_name, up_to=9999)
Executes schema migrations.
migrations_package_name
- fully qualified name of the Python package containing the migrations.up_to
- number of the last migration to apply.
paginate(model_class, order_by, page_num=1, page_size=100, conditions=None, settings=None)
Selects records and returns a single page of model instances.
model_class
: the model class matching the query's table, orNone
for getting back instances of an ad-hoc model.order_by
: columns to use for sorting the query (contents of the ORDER BY clause).page_num
: the page number (1-based), or -1 to get the last page.page_size
: number of records to return per page.conditions
: optional SQL conditions (contents of the WHERE clause).settings
: query settings to send as HTTP GET parameters
The result is a namedtuple containing objects
(list), number_of_objects
,
pages_total
, number
(of the current page), and page_size
.
raw(query, settings=None, stream=False)
Performs a query and returns its output as text.
query
: the SQL query to execute.settings
: query settings to send as HTTP GET parametersstream
: if true, the HTTP response from ClickHouse will be streamed.
select(query, model_class=None, settings=None)
Performs a query and returns a generator of model instances.
query
: the SQL query to execute.model_class
: the model class matching the query's table, orNone
for getting back instances of an ad-hoc model.settings
: query settings to send as HTTP GET parameters
DatabaseException
Extends Exception
Raised when a database operation fails.
infi.clickhouse_orm.models
Model
A base class for ORM models. Each model class represent a ClickHouse table. For example:
class CPUStats(Model):
timestamp = DateTimeField()
cpu_id = UInt16Field()
cpu_percent = Float32Field()
engine = Memory()
Model(**kwargs)
Creates a model instance, using keyword arguments as field values.
Since values are immediately converted to their Pythonic type,
invalid values will cause a ValueError
to be raised.
Unrecognized field names will cause an AttributeError
.
Model.create_table_sql(db)
Returns the SQL command for creating a table for this model.
Model.drop_table_sql(db)
Returns the SQL command for deleting this model's table.
Model.fields(writable=False)
Returns an OrderedDict
of the model's fields (from name to Field
instance).
If writable
is true, only writable fields are included.
Callers should not modify the dictionary.
Model.from_tsv(line, field_names, timezone_in_use=UTC, database=None)
Create a model instance from a tab-separated line. The line may or may not include a newline.
The field_names
list must match the fields defined in the model, but does not have to include all of them.
line
: the TSV-formatted data.field_names
: names of the model fields in the data.timezone_in_use
: the timezone to use when parsing dates and datetimes.database
: if given, sets the database that this instance belongs to.
get_database()
Gets the Database
that this model instance belongs to.
Returns None
unless the instance was read from the database or written to it.
get_field(name)
Gets a Field
instance given its name, or None
if not found.
Model.is_read_only()
Returns true if the model is marked as read only.
Model.is_system_model()
Returns true if the model represents a system table.
Model.objects_in(database)
Returns a QuerySet
for selecting instances of this model class.
set_database(db)
Sets the Database
that this model instance belongs to.
This is done automatically when the instance is read from the database or written to it.
Model.table_name()
Returns the model's database table name. By default this is the class name converted to lowercase. Override this if you want to use a different table name.
to_dict(include_readonly=True, field_names=None)
Returns the instance's column values as a dict.
include_readonly
: if false, returns only fields that can be inserted into database.field_names
: an iterable of field names to return (optional)
to_tsv(include_readonly=True)
Returns the instance's column values as a tab-separated line. A newline is not included.
include_readonly
: if false, returns only fields that can be inserted into database.
BufferModel
Extends Model
BufferModel(**kwargs)
Creates a model instance, using keyword arguments as field values.
Since values are immediately converted to their Pythonic type,
invalid values will cause a ValueError
to be raised.
Unrecognized field names will cause an AttributeError
.
BufferModel.create_table_sql(db)
Returns the SQL command for creating a table for this model.
BufferModel.drop_table_sql(db)
Returns the SQL command for deleting this model's table.
BufferModel.fields(writable=False)
Returns an OrderedDict
of the model's fields (from name to Field
instance).
If writable
is true, only writable fields are included.
Callers should not modify the dictionary.
BufferModel.from_tsv(line, field_names, timezone_in_use=UTC, database=None)
Create a model instance from a tab-separated line. The line may or may not include a newline.
The field_names
list must match the fields defined in the model, but does not have to include all of them.
line
: the TSV-formatted data.field_names
: names of the model fields in the data.timezone_in_use
: the timezone to use when parsing dates and datetimes.database
: if given, sets the database that this instance belongs to.
get_database()
Gets the Database
that this model instance belongs to.
Returns None
unless the instance was read from the database or written to it.
get_field(name)
Gets a Field
instance given its name, or None
if not found.
BufferModel.is_read_only()
Returns true if the model is marked as read only.
BufferModel.is_system_model()
Returns true if the model represents a system table.
BufferModel.objects_in(database)
Returns a QuerySet
for selecting instances of this model class.
set_database(db)
Sets the Database
that this model instance belongs to.
This is done automatically when the instance is read from the database or written to it.
BufferModel.table_name()
Returns the model's database table name. By default this is the class name converted to lowercase. Override this if you want to use a different table name.
to_dict(include_readonly=True, field_names=None)
Returns the instance's column values as a dict.
include_readonly
: if false, returns only fields that can be inserted into database.field_names
: an iterable of field names to return (optional)
to_tsv(include_readonly=True)
Returns the instance's column values as a tab-separated line. A newline is not included.
include_readonly
: if false, returns only fields that can be inserted into database.
DistributedModel
Extends Model
Model for Distributed engine
DistributedModel(**kwargs)
Creates a model instance, using keyword arguments as field values.
Since values are immediately converted to their Pythonic type,
invalid values will cause a ValueError
to be raised.
Unrecognized field names will cause an AttributeError
.
DistributedModel.create_table_sql(db)
DistributedModel.drop_table_sql(db)
Returns the SQL command for deleting this model's table.
DistributedModel.fields(writable=False)
Returns an OrderedDict
of the model's fields (from name to Field
instance).
If writable
is true, only writable fields are included.
Callers should not modify the dictionary.
DistributedModel.fix_engine_table()
Remember: Distributed table does not store any data, just provides distributed access to it.
So if we define a model with engine that has no defined table for data storage (see FooDistributed below), that table cannot be successfully created. This routine can automatically fix engine's storage table by finding the first non-distributed model among your model's superclasses.
class Foo(Model): ... id = UInt8Field(1) ... class FooDistributed(Foo, DistributedModel): ... engine = Distributed('my_cluster') ... FooDistributed.engine.table None FooDistributed.fix_engine() FooDistributed.engine.table <class 'main.Foo'>
However if you prefer more explicit way of doing things, you can always mention the Foo model twice without bothering with any fixes:
class FooDistributedVerbose(Foo, DistributedModel): ... engine = Distributed('my_cluster', Foo) FooDistributedVerbose.engine.table <class 'main.Foo'>
See tests.test_engines:DistributedTestCase for more examples
DistributedModel.from_tsv(line, field_names, timezone_in_use=UTC, database=None)
Create a model instance from a tab-separated line. The line may or may not include a newline.
The field_names
list must match the fields defined in the model, but does not have to include all of them.
line
: the TSV-formatted data.field_names
: names of the model fields in the data.timezone_in_use
: the timezone to use when parsing dates and datetimes.database
: if given, sets the database that this instance belongs to.
get_database()
Gets the Database
that this model instance belongs to.
Returns None
unless the instance was read from the database or written to it.
get_field(name)
Gets a Field
instance given its name, or None
if not found.
DistributedModel.is_read_only()
Returns true if the model is marked as read only.
DistributedModel.is_system_model()
Returns true if the model represents a system table.
DistributedModel.objects_in(database)
Returns a QuerySet
for selecting instances of this model class.
set_database(db)
DistributedModel.table_name()
Returns the model's database table name. By default this is the class name converted to lowercase. Override this if you want to use a different table name.
to_dict(include_readonly=True, field_names=None)
Returns the instance's column values as a dict.
include_readonly
: if false, returns only fields that can be inserted into database.field_names
: an iterable of field names to return (optional)
to_tsv(include_readonly=True)
Returns the instance's column values as a tab-separated line. A newline is not included.
include_readonly
: if false, returns only fields that can be inserted into database.
infi.clickhouse_orm.fields
ArrayField
Extends Field
ArrayField(inner_field, default=None, alias=None, materialized=None, readonly=None, codec=None)
BaseEnumField
Extends Field
Abstract base class for all enum-type fields.
BaseEnumField(enum_cls, default=None, alias=None, materialized=None, readonly=None, codec=None)
BaseFloatField
Extends Field
Abstract base class for all float-type fields.
BaseFloatField(default=None, alias=None, materialized=None, readonly=None, codec=None)
BaseIntField
Extends Field
Abstract base class for all integer-type fields.
BaseIntField(default=None, alias=None, materialized=None, readonly=None, codec=None)
DateField
Extends Field
DateField(default=None, alias=None, materialized=None, readonly=None, codec=None)
DateTimeField
Extends Field
DateTimeField(default=None, alias=None, materialized=None, readonly=None, codec=None)
Decimal128Field
Extends DecimalField
Decimal128Field(scale, default=None, alias=None, materialized=None, readonly=None)
Decimal32Field
Extends DecimalField
Decimal32Field(scale, default=None, alias=None, materialized=None, readonly=None)
Decimal64Field
Extends DecimalField
Decimal64Field(scale, default=None, alias=None, materialized=None, readonly=None)
DecimalField
Extends Field
Base class for all decimal fields. Can also be used directly.
DecimalField(precision, scale, default=None, alias=None, materialized=None, readonly=None)
Enum16Field
Extends BaseEnumField
Enum16Field(enum_cls, default=None, alias=None, materialized=None, readonly=None, codec=None)
Enum8Field
Extends BaseEnumField
Enum8Field(enum_cls, default=None, alias=None, materialized=None, readonly=None, codec=None)
Field
Abstract base class for all field types.
Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
FixedStringField
Extends StringField
FixedStringField(length, default=None, alias=None, materialized=None, readonly=None)
Float32Field
Extends BaseFloatField
Float32Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
Float64Field
Extends BaseFloatField
Float64Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
Int16Field
Extends BaseIntField
Int16Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
Int32Field
Extends BaseIntField
Int32Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
Int64Field
Extends BaseIntField
Int64Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
Int8Field
Extends BaseIntField
Int8Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
LowCardinalityField
Extends Field
LowCardinalityField(inner_field, default=None, alias=None, materialized=None, readonly=None, codec=None)
NullableField
Extends Field
NullableField(inner_field, default=None, alias=None, materialized=None, extra_null_values=None, codec=None)
StringField
Extends Field
StringField(default=None, alias=None, materialized=None, readonly=None, codec=None)
UInt16Field
Extends BaseIntField
UInt16Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
UInt32Field
Extends BaseIntField
UInt32Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
UInt64Field
Extends BaseIntField
UInt64Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
UInt8Field
Extends BaseIntField
UInt8Field(default=None, alias=None, materialized=None, readonly=None, codec=None)
UUIDField
Extends Field
UUIDField(default=None, alias=None, materialized=None, readonly=None, codec=None)
infi.clickhouse_orm.engines
Engine
TinyLog
Extends Engine
Log
Extends Engine
Memory
Extends Engine
MergeTree
Extends Engine
MergeTree(date_col=None, order_by=(), sampling_expr=None, index_granularity=8192, replica_table_path=None, replica_name=None, partition_key=None)
Buffer
Extends Engine
Buffers the data to write in RAM, periodically flushing it to another table.
Must be used in conjuction with a BufferModel
.
Read more here.
Buffer(main_model, num_layers=16, min_time=10, max_time=100, min_rows=10000, max_rows=1000000, min_bytes=10000000, max_bytes=100000000)
Merge
Extends Engine
The Merge engine (not to be confused with MergeTree) does not store data itself, but allows reading from any number of other tables simultaneously. Writing to a table is not supported https://clickhouse.yandex/docs/en/single/index.html#document-table_engines/merge
Merge(table_regex)
Distributed
Extends Engine
The Distributed engine by itself does not store data, but allows distributed query processing on multiple servers. Reading is automatically parallelized. During a read, the table indexes on remote servers are used, if there are any.
See full documentation here https://clickhouse.yandex/docs/en/table_engines/distributed.html
Distributed(cluster, table=None, sharding_key=None)
:param cluster: what cluster to access data from :param table: underlying table that actually stores data. If you are not specifying any table here, ensure that it can be inferred from your model's superclass (see models.DistributedModel.fix_engine_table) :param sharding_key: how to distribute data among shards when inserting straightly into Distributed table, optional
CollapsingMergeTree
Extends MergeTree
CollapsingMergeTree(date_col=None, order_by=(), sign_col="sign", sampling_expr=None, index_granularity=8192, replica_table_path=None, replica_name=None, partition_key=None)
SummingMergeTree
Extends MergeTree
SummingMergeTree(date_col=None, order_by=(), summing_cols=None, sampling_expr=None, index_granularity=8192, replica_table_path=None, replica_name=None, partition_key=None)
ReplacingMergeTree
Extends MergeTree
ReplacingMergeTree(date_col=None, order_by=(), ver_col=None, sampling_expr=None, index_granularity=8192, replica_table_path=None, replica_name=None, partition_key=None)
infi.clickhouse_orm.query
QuerySet
A queryset is an object that represents a database query using a specific Model
.
It is lazy, meaning that it does not hit the database until you iterate over its
matching rows (model instances).
QuerySet(model_cls, database)
Initializer. It is possible to create a queryset like this, but the standard
way is to use MyModel.objects_in(database)
.
aggregate(*args, **kwargs)
Returns an AggregateQuerySet
over this query, with args
serving as
grouping fields and kwargs
serving as calculated fields. At least one
calculated field is required. For example:
Event.objects_in(database).filter(date__gt='2017-08-01').aggregate('event_type', count='count()')
is equivalent to:
SELECT event_type, count() AS count FROM event
WHERE data > '2017-08-01'
GROUP BY event_type
as_sql()
Returns the whole query as a SQL string.
conditions_as_sql(prewhere=False)
Returns the contents of the query's WHERE
or PREWHERE
clause as a string.
count()
Returns the number of matching model instances.
distinct()
Adds a DISTINCT clause to the query, meaning that any duplicate rows in the results will be omitted.
exclude(*q, **kwargs)
Returns a copy of this queryset that excludes all rows matching the conditions.
Pass prewhere=True
to apply the conditions as PREWHERE instead of WHERE.
filter(*q, **kwargs)
Returns a copy of this queryset that includes only rows matching the conditions.
Pass prewhere=True
to apply the conditions as PREWHERE instead of WHERE.
final()
Adds a FINAL modifier to table, meaning data will be collapsed to final version.
Can be used with CollapsingMergeTree
engine only.
only(*field_names)
Returns a copy of this queryset limited to the specified field names. Useful when there are large fields that are not needed, or for creating a subquery to use with an IN operator.
order_by(*field_names)
Returns a copy of this queryset with the ordering changed.
order_by_as_sql()
Returns the contents of the query's ORDER BY
clause as a string.
paginate(page_num=1, page_size=100)
Returns a single page of model instances that match the queryset.
Note that order_by
should be used first, to ensure a correct
partitioning of records into pages.
page_num
: the page number (1-based), or -1 to get the last page.page_size
: number of records to return per page.
The result is a namedtuple containing objects
(list), number_of_objects
,
pages_total
, number
(of the current page), and page_size
.
select_fields_as_sql()
Returns the selected fields or expressions as a SQL string.
AggregateQuerySet
Extends QuerySet
A queryset used for aggregation.
AggregateQuerySet(base_qs, grouping_fields, calculated_fields)
Initializer. Normally you should not call this but rather use QuerySet.aggregate()
.
The grouping fields should be a list/tuple of field names from the model. For example:
('event_type', 'event_subtype')
The calculated fields should be a mapping from name to a ClickHouse aggregation function. For example:
{'weekday': 'toDayOfWeek(event_date)', 'number_of_events': 'count()'}
At least one calculated field is required.
aggregate(*args, **kwargs)
This method is not supported on AggregateQuerySet
.
as_sql()
Returns the whole query as a SQL string.
conditions_as_sql(prewhere=False)
Returns the contents of the query's WHERE
or PREWHERE
clause as a string.
count()
Returns the number of rows after aggregation.
distinct()
Adds a DISTINCT clause to the query, meaning that any duplicate rows in the results will be omitted.
exclude(*q, **kwargs)
Returns a copy of this queryset that excludes all rows matching the conditions.
Pass prewhere=True
to apply the conditions as PREWHERE instead of WHERE.
filter(*q, **kwargs)
Returns a copy of this queryset that includes only rows matching the conditions.
Pass prewhere=True
to apply the conditions as PREWHERE instead of WHERE.
final()
Adds a FINAL modifier to table, meaning data will be collapsed to final version.
Can be used with CollapsingMergeTree
engine only.
group_by(*args)
This method lets you specify the grouping fields explicitly. The args
must
be names of grouping fields or calculated fields that this queryset was
created with.
only(*field_names)
This method is not supported on AggregateQuerySet
.
order_by(*field_names)
Returns a copy of this queryset with the ordering changed.
order_by_as_sql()
Returns the contents of the query's ORDER BY
clause as a string.
paginate(page_num=1, page_size=100)
Returns a single page of model instances that match the queryset.
Note that order_by
should be used first, to ensure a correct
partitioning of records into pages.
page_num
: the page number (1-based), or -1 to get the last page.page_size
: number of records to return per page.
The result is a namedtuple containing objects
(list), number_of_objects
,
pages_total
, number
(of the current page), and page_size
.
select_fields_as_sql()
Returns the selected fields or expressions as a SQL string.
with_totals()
Adds WITH TOTALS modifier ot GROUP BY, making query return extra row with aggregate function calculated across all the rows. More information: https://clickhouse.yandex/docs/en/query_language/select/#with-totals-modifier