2018-06-09 22:22:54 +03:00
|
|
|
import abc
|
2018-06-14 20:35:12 +03:00
|
|
|
import asyncio
|
2017-06-08 14:12:57 +03:00
|
|
|
import logging
|
2018-03-02 22:05:09 +03:00
|
|
|
import platform
|
2018-06-18 14:22:25 +03:00
|
|
|
import time
|
2019-01-24 13:16:40 +03:00
|
|
|
from datetime import datetime, timezone
|
2018-06-08 22:52:59 +03:00
|
|
|
|
2019-03-21 14:21:00 +03:00
|
|
|
from .. import version, helpers, __name__ as __base_name__
|
2018-06-09 22:22:54 +03:00
|
|
|
from ..crypto import rsa
|
|
|
|
from ..extensions import markdown
|
2019-03-10 03:00:11 +03:00
|
|
|
from ..network import MTProtoSender, ConnectionTcpFull, TcpMTProxy
|
2018-08-02 16:17:44 +03:00
|
|
|
from ..sessions import Session, SQLiteSession, MemorySession
|
2018-06-20 20:48:00 +03:00
|
|
|
from ..tl import TLObject, functions, types
|
2018-06-18 22:02:42 +03:00
|
|
|
from ..tl.alltlobjects import LAYER
|
2019-03-26 13:27:21 +03:00
|
|
|
from ..entitycache import EntityCache
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2017-12-28 03:04:11 +03:00
|
|
|
DEFAULT_DC_ID = 4
|
2017-11-16 15:30:18 +03:00
|
|
|
DEFAULT_IPV4_IP = '149.154.167.51'
|
|
|
|
DEFAULT_IPV6_IP = '[2001:67c:4e8:f002::a]'
|
|
|
|
DEFAULT_PORT = 443
|
|
|
|
|
2019-01-11 17:52:30 +03:00
|
|
|
__default_log__ = logging.getLogger(__base_name__)
|
|
|
|
__default_log__.addHandler(logging.NullHandler())
|
2017-12-20 14:47:10 +03:00
|
|
|
|
2017-11-16 15:30:18 +03:00
|
|
|
|
2019-02-27 21:30:12 +03:00
|
|
|
# TODO How hard would it be to support both `trio` and `asyncio`?
|
2018-06-09 22:22:54 +03:00
|
|
|
class TelegramBaseClient(abc.ABC):
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
2018-06-09 22:22:54 +03:00
|
|
|
This is the abstract base class for the client. It defines some
|
|
|
|
basic stuff like connecting, switching data center, etc, and
|
|
|
|
leaves the `__call__` unimplemented.
|
2018-06-08 22:52:59 +03:00
|
|
|
|
|
|
|
Args:
|
|
|
|
session (`str` | `telethon.sessions.abstract.Session`, `None`):
|
|
|
|
The file name of the session file to be used if a string is
|
|
|
|
given (it may be a full path), or the Session instance to be
|
|
|
|
used otherwise. If it's ``None``, the session will not be saved,
|
|
|
|
and you should call :meth:`.log_out()` when you're done.
|
|
|
|
|
|
|
|
Note that if you pass a string it will be a file in the current
|
|
|
|
working directory, although you can also pass absolute paths.
|
|
|
|
|
|
|
|
The session file contains enough information for you to login
|
|
|
|
without re-sending the code, so if you have to enter the code
|
|
|
|
more than once, maybe you're changing the working directory,
|
|
|
|
renaming or removing the file, or using random names.
|
|
|
|
|
|
|
|
api_id (`int` | `str`):
|
|
|
|
The API ID you obtained from https://my.telegram.org.
|
|
|
|
|
|
|
|
api_hash (`str`):
|
|
|
|
The API ID you obtained from https://my.telegram.org.
|
|
|
|
|
|
|
|
connection (`telethon.network.connection.common.Connection`, optional):
|
|
|
|
The connection instance to be used when creating a new connection
|
2018-09-28 18:51:28 +03:00
|
|
|
to the servers. It **must** be a type.
|
2018-06-08 22:52:59 +03:00
|
|
|
|
|
|
|
Defaults to `telethon.network.connection.tcpfull.ConnectionTcpFull`.
|
|
|
|
|
|
|
|
use_ipv6 (`bool`, optional):
|
|
|
|
Whether to connect to the servers through IPv6 or not.
|
|
|
|
By default this is ``False`` as IPv6 support is not
|
|
|
|
too widespread yet.
|
|
|
|
|
2019-02-11 11:54:35 +03:00
|
|
|
proxy (`tuple` | `list` | `dict`, optional):
|
|
|
|
An iterable consisting of the proxy info. If `connection` is
|
2019-03-10 03:00:11 +03:00
|
|
|
one of `MTProxy`, then it should contain MTProxy credentials:
|
2019-02-11 11:54:35 +03:00
|
|
|
``('hostname', port, 'secret')``. Otherwise, it's meant to store
|
|
|
|
function parameters for PySocks, like ``(type, 'hostname', port)``.
|
2018-06-08 22:52:59 +03:00
|
|
|
See https://github.com/Anorov/PySocks#usage-1 for more.
|
|
|
|
|
2018-10-05 15:06:15 +03:00
|
|
|
timeout (`int` | `float`, optional):
|
|
|
|
The timeout in seconds to be used when connecting.
|
|
|
|
This is **not** the timeout to be used when ``await``'ing for
|
|
|
|
invoked requests, and you should use ``asyncio.wait`` or
|
|
|
|
``asyncio.wait_for`` for that.
|
2018-06-18 19:11:16 +03:00
|
|
|
|
2019-02-06 21:41:45 +03:00
|
|
|
request_retries (`int` | `None`, optional):
|
2018-06-18 19:11:16 +03:00
|
|
|
How many times a request should be retried. Request are retried
|
|
|
|
when Telegram is having internal issues (due to either
|
|
|
|
``errors.ServerError`` or ``errors.RpcCallFailError``),
|
|
|
|
when there is a ``errors.FloodWaitError`` less than
|
2018-06-26 17:09:16 +03:00
|
|
|
`flood_sleep_threshold`, or when there's a migrate error.
|
2018-06-18 19:11:16 +03:00
|
|
|
|
2019-02-06 21:41:45 +03:00
|
|
|
May take a negative or ``None`` value for infinite retries, but
|
|
|
|
this is not recommended, since some requests can always trigger
|
|
|
|
a call fail (such as searching for messages).
|
2018-06-18 19:11:16 +03:00
|
|
|
|
2019-02-06 21:41:45 +03:00
|
|
|
connection_retries (`int` | `None`, optional):
|
2018-06-18 19:11:16 +03:00
|
|
|
How many times the reconnection should retry, either on the
|
|
|
|
initial connection or when Telegram disconnects us. May be
|
2019-02-06 21:41:45 +03:00
|
|
|
set to a negative or ``None`` value for infinite retries, but
|
|
|
|
this is not recommended, since the program can get stuck in an
|
|
|
|
infinite loop.
|
2018-06-18 19:11:16 +03:00
|
|
|
|
2018-10-28 12:55:58 +03:00
|
|
|
retry_delay (`int` | `float`, optional):
|
|
|
|
The delay in seconds to sleep between automatic reconnections.
|
|
|
|
|
2018-06-18 19:11:16 +03:00
|
|
|
auto_reconnect (`bool`, optional):
|
|
|
|
Whether reconnection should be retried `connection_retries`
|
|
|
|
times automatically if Telegram disconnects us or not.
|
2018-06-08 22:52:59 +03:00
|
|
|
|
2018-06-29 11:45:04 +03:00
|
|
|
sequential_updates (`bool`, optional):
|
|
|
|
By default every incoming update will create a new task, so
|
|
|
|
you can handle several updates in parallel. Some scripts need
|
|
|
|
the order in which updates are processed to be sequential, and
|
|
|
|
this setting allows them to do so.
|
|
|
|
|
|
|
|
If set to ``True``, incoming updates will be put in a queue
|
|
|
|
and processed sequentially. This means your event handlers
|
|
|
|
should *not* perform long-running operations since new
|
|
|
|
updates are put inside of an unbounded queue.
|
|
|
|
|
2018-06-26 17:09:16 +03:00
|
|
|
flood_sleep_threshold (`int` | `float`, optional):
|
|
|
|
The threshold below which the library should automatically
|
|
|
|
sleep on flood wait errors (inclusive). For instance, if a
|
|
|
|
``FloodWaitError`` for 17s occurs and `flood_sleep_threshold`
|
|
|
|
is 20s, the library will ``sleep`` automatically. If the error
|
|
|
|
was for 21s, it would ``raise FloodWaitError`` instead. Values
|
|
|
|
larger than a day (like ``float('inf')``) will be changed to a day.
|
2018-06-08 22:52:59 +03:00
|
|
|
|
|
|
|
device_model (`str`, optional):
|
|
|
|
"Device model" to be sent when creating the initial connection.
|
|
|
|
Defaults to ``platform.node()``.
|
|
|
|
|
|
|
|
system_version (`str`, optional):
|
|
|
|
"System version" to be sent when creating the initial connection.
|
|
|
|
Defaults to ``platform.system()``.
|
|
|
|
|
|
|
|
app_version (`str`, optional):
|
|
|
|
"App version" to be sent when creating the initial connection.
|
|
|
|
Defaults to `telethon.version.__version__`.
|
|
|
|
|
|
|
|
lang_code (`str`, optional):
|
|
|
|
"Language code" to be sent when creating the initial connection.
|
|
|
|
Defaults to ``'en'``.
|
|
|
|
|
|
|
|
system_lang_code (`str`, optional):
|
|
|
|
"System lang code" to be sent when creating the initial connection.
|
|
|
|
Defaults to `lang_code`.
|
2019-01-11 17:52:30 +03:00
|
|
|
|
|
|
|
loop (`asyncio.AbstractEventLoop`, optional):
|
|
|
|
Asyncio event loop to use. Defaults to `asyncio.get_event_loop()`
|
|
|
|
|
|
|
|
base_logger (`str` | `logging.Logger`, optional):
|
|
|
|
Base logger name or instance to use.
|
|
|
|
If a `str` is given, it'll be passed to `logging.getLogger()`. If a
|
|
|
|
`logging.Logger` is given, it'll be used directly. If something
|
|
|
|
else or nothing is given, the default logger will be used.
|
2017-06-08 14:12:57 +03:00
|
|
|
"""
|
|
|
|
|
|
|
|
# Current TelegramClient version
|
2017-10-28 13:21:07 +03:00
|
|
|
__version__ = version.__version__
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-06-10 22:30:16 +03:00
|
|
|
# Cached server configuration (with .dc_options), can be "global"
|
2018-06-08 22:52:59 +03:00
|
|
|
_config = None
|
2018-06-10 22:30:16 +03:00
|
|
|
_cdn_config = None
|
2017-09-17 15:30:23 +03:00
|
|
|
|
2017-06-08 14:12:57 +03:00
|
|
|
# region Initialization
|
|
|
|
|
2017-06-22 12:43:42 +03:00
|
|
|
def __init__(self, session, api_id, api_hash,
|
2018-05-10 15:22:19 +03:00
|
|
|
*,
|
|
|
|
connection=ConnectionTcpFull,
|
2017-11-16 15:30:18 +03:00
|
|
|
use_ipv6=False,
|
2017-09-07 19:49:08 +03:00
|
|
|
proxy=None,
|
2018-10-05 15:06:15 +03:00
|
|
|
timeout=10,
|
2018-06-18 19:11:16 +03:00
|
|
|
request_retries=5,
|
|
|
|
connection_retries=5,
|
2018-10-28 12:55:58 +03:00
|
|
|
retry_delay=1,
|
2018-06-18 19:11:16 +03:00
|
|
|
auto_reconnect=True,
|
2018-06-29 11:45:04 +03:00
|
|
|
sequential_updates=False,
|
2018-06-26 17:09:16 +03:00
|
|
|
flood_sleep_threshold=60,
|
2018-03-02 22:05:09 +03:00
|
|
|
device_model=None,
|
|
|
|
system_version=None,
|
|
|
|
app_version=None,
|
|
|
|
lang_code='en',
|
2018-06-14 20:35:12 +03:00
|
|
|
system_lang_code='en',
|
2019-01-11 17:52:30 +03:00
|
|
|
loop=None,
|
|
|
|
base_logger=None):
|
2017-09-29 21:50:27 +03:00
|
|
|
if not api_id or not api_hash:
|
2017-12-28 02:22:28 +03:00
|
|
|
raise ValueError(
|
2017-09-29 21:50:27 +03:00
|
|
|
"Your API ID or Hash cannot be empty or None. "
|
2018-01-08 16:04:04 +03:00
|
|
|
"Refer to telethon.rtfd.io for more information.")
|
2017-09-29 21:50:27 +03:00
|
|
|
|
2017-11-16 15:30:18 +03:00
|
|
|
self._use_ipv6 = use_ipv6
|
2018-06-14 20:35:12 +03:00
|
|
|
self._loop = loop or asyncio.get_event_loop()
|
2018-03-02 00:34:32 +03:00
|
|
|
|
2019-01-11 17:52:30 +03:00
|
|
|
if isinstance(base_logger, str):
|
|
|
|
base_logger = logging.getLogger(base_logger)
|
|
|
|
elif not isinstance(base_logger, logging.Logger):
|
|
|
|
base_logger = __default_log__
|
|
|
|
|
|
|
|
class _Loggers(dict):
|
|
|
|
def __missing__(self, key):
|
|
|
|
if key.startswith("telethon."):
|
2019-01-12 14:15:29 +03:00
|
|
|
key = key.split('.', maxsplit=1)[1]
|
|
|
|
|
2019-01-11 17:52:30 +03:00
|
|
|
return base_logger.getChild(key)
|
|
|
|
|
|
|
|
self._log = _Loggers()
|
|
|
|
|
2017-09-29 21:50:27 +03:00
|
|
|
# Determine what session object we have
|
|
|
|
if isinstance(session, str) or session is None:
|
2018-08-02 16:17:44 +03:00
|
|
|
try:
|
|
|
|
session = SQLiteSession(session)
|
2019-02-13 10:51:26 +03:00
|
|
|
except ImportError:
|
2018-08-02 16:17:44 +03:00
|
|
|
import warnings
|
|
|
|
warnings.warn(
|
|
|
|
'The sqlite3 module is not available under this '
|
|
|
|
'Python installation and no custom session '
|
|
|
|
'instance was given; using MemorySession.\n'
|
|
|
|
'You will need to re-login every time unless '
|
|
|
|
'you use another session storage'
|
|
|
|
)
|
|
|
|
session = MemorySession()
|
2017-09-29 21:50:27 +03:00
|
|
|
elif not isinstance(session, Session):
|
2017-12-28 02:22:28 +03:00
|
|
|
raise TypeError(
|
2017-09-29 21:50:27 +03:00
|
|
|
'The given session must be a str or a Session instance.'
|
|
|
|
)
|
|
|
|
|
2017-11-16 15:40:25 +03:00
|
|
|
# ':' in session.server_address is True if it's an IPv6 address
|
|
|
|
if (not session.server_address or
|
|
|
|
(':' in session.server_address) != use_ipv6):
|
2017-12-28 03:04:11 +03:00
|
|
|
session.set_dc(
|
|
|
|
DEFAULT_DC_ID,
|
|
|
|
DEFAULT_IPV6_IP if self._use_ipv6 else DEFAULT_IPV4_IP,
|
|
|
|
DEFAULT_PORT
|
|
|
|
)
|
2017-11-16 15:30:18 +03:00
|
|
|
|
2018-06-26 17:09:16 +03:00
|
|
|
self.flood_sleep_threshold = flood_sleep_threshold
|
2018-10-12 23:00:02 +03:00
|
|
|
|
2019-03-26 13:27:21 +03:00
|
|
|
# TODO Use AsyncClassWrapper(session)
|
|
|
|
# ChatGetter and SenderGetter can use the in-memory _entity_cache
|
|
|
|
# to avoid network access and the need for await in session files.
|
2018-10-12 23:00:02 +03:00
|
|
|
#
|
2019-03-26 13:27:21 +03:00
|
|
|
# The session files only wants the entities to persist
|
|
|
|
# them to disk, and to save additional useful information.
|
|
|
|
# TODO Session should probably return all cached
|
|
|
|
# info of entities, not just the input versions
|
2018-10-12 23:00:02 +03:00
|
|
|
self.session = session
|
2019-03-26 13:27:21 +03:00
|
|
|
self._entity_cache = EntityCache()
|
2017-06-11 23:42:04 +03:00
|
|
|
self.api_id = int(api_id)
|
2017-06-08 14:12:57 +03:00
|
|
|
self.api_hash = api_hash
|
2017-09-21 14:43:33 +03:00
|
|
|
|
2019-02-06 21:41:45 +03:00
|
|
|
self._request_retries = request_retries
|
|
|
|
self._connection_retries = connection_retries
|
2018-10-28 12:55:58 +03:00
|
|
|
self._retry_delay = retry_delay or 0
|
2018-10-04 18:11:31 +03:00
|
|
|
self._proxy = proxy
|
2018-10-04 17:39:57 +03:00
|
|
|
self._timeout = timeout
|
2018-06-18 19:11:16 +03:00
|
|
|
self._auto_reconnect = auto_reconnect
|
|
|
|
|
2018-09-28 18:51:28 +03:00
|
|
|
assert isinstance(connection, type)
|
|
|
|
self._connection = connection
|
2019-03-10 03:00:11 +03:00
|
|
|
init_proxy = None if not issubclass(connection, TcpMTProxy) else \
|
|
|
|
types.InputClientProxy(*connection.address_info(proxy))
|
2018-05-10 15:22:19 +03:00
|
|
|
|
2018-06-11 21:05:10 +03:00
|
|
|
# Used on connection. Capture the variables in a lambda since
|
|
|
|
# exporting clients need to create this InvokeWithLayerRequest.
|
2018-06-09 22:03:48 +03:00
|
|
|
system = platform.uname()
|
2018-06-11 21:05:10 +03:00
|
|
|
self._init_with = lambda x: functions.InvokeWithLayerRequest(
|
2018-06-09 22:03:48 +03:00
|
|
|
LAYER, functions.InitConnectionRequest(
|
|
|
|
api_id=self.api_id,
|
|
|
|
device_model=device_model or system.system or 'Unknown',
|
|
|
|
system_version=system_version or system.release or '1.0',
|
|
|
|
app_version=app_version or self.__version__,
|
|
|
|
lang_code=lang_code,
|
|
|
|
system_lang_code=system_lang_code,
|
|
|
|
lang_pack='', # "langPacks are for official apps only"
|
2019-02-11 02:16:46 +03:00
|
|
|
query=x,
|
|
|
|
proxy=init_proxy
|
2018-06-09 22:03:48 +03:00
|
|
|
)
|
|
|
|
)
|
2017-09-22 13:20:38 +03:00
|
|
|
|
2018-06-11 21:05:10 +03:00
|
|
|
self._sender = MTProtoSender(
|
2018-10-19 14:50:11 +03:00
|
|
|
self.session.auth_key, self._loop,
|
2019-01-11 17:52:30 +03:00
|
|
|
loggers=self._log,
|
2018-06-18 19:11:16 +03:00
|
|
|
retries=self._connection_retries,
|
2018-10-28 12:55:58 +03:00
|
|
|
delay=self._retry_delay,
|
2018-06-18 19:11:16 +03:00
|
|
|
auto_reconnect=self._auto_reconnect,
|
2018-10-04 17:39:57 +03:00
|
|
|
connect_timeout=self._timeout,
|
2018-11-03 20:53:26 +03:00
|
|
|
auth_key_callback=self._auth_key_callback,
|
2018-06-27 11:15:59 +03:00
|
|
|
update_callback=self._handle_update,
|
2018-06-27 20:40:32 +03:00
|
|
|
auto_reconnect_callback=self._handle_auto_reconnect
|
2018-06-11 21:05:10 +03:00
|
|
|
)
|
|
|
|
|
2018-07-21 13:25:20 +03:00
|
|
|
# Remember flood-waited requests to avoid making them again
|
|
|
|
self._flood_waited_requests = {}
|
|
|
|
|
2018-06-28 15:10:36 +03:00
|
|
|
# Cache ``{dc_id: (n, MTProtoSender)}`` for all borrowed senders,
|
|
|
|
# being ``n`` the amount of borrows a given sender has; once ``n``
|
|
|
|
# reaches ``0`` it should be disconnected and removed.
|
|
|
|
self._borrowed_senders = {}
|
2018-08-21 12:31:14 +03:00
|
|
|
self._borrow_sender_lock = asyncio.Lock(loop=self._loop)
|
2017-07-04 11:21:15 +03:00
|
|
|
|
2018-06-18 14:22:25 +03:00
|
|
|
self._updates_handle = None
|
|
|
|
self._last_request = time.time()
|
2018-06-20 20:48:00 +03:00
|
|
|
self._channel_pts = {}
|
2018-06-21 10:32:09 +03:00
|
|
|
|
2018-06-29 11:45:04 +03:00
|
|
|
if sequential_updates:
|
2018-08-21 12:31:14 +03:00
|
|
|
self._updates_queue = asyncio.Queue(loop=self._loop)
|
|
|
|
self._dispatching_updates_queue = asyncio.Event(loop=self._loop)
|
2018-06-29 11:45:04 +03:00
|
|
|
else:
|
|
|
|
self._updates_queue = None
|
|
|
|
self._dispatching_updates_queue = None
|
|
|
|
|
2018-12-06 18:07:11 +03:00
|
|
|
self._authorized = None # None = unknown, False = no, True = yes
|
2019-03-28 14:32:02 +03:00
|
|
|
|
|
|
|
# Update state (for catching up after a disconnection)
|
|
|
|
#
|
2019-04-10 20:09:15 +03:00
|
|
|
# We only care about the pts and the date. By using a tuple which
|
|
|
|
# is lightweight and immutable we can easily copy them around to
|
|
|
|
# each update in case they need to fetch missing entities.
|
|
|
|
state = self.session.get_update_state(0)
|
|
|
|
self._old_pts_date = state.pts, state.date
|
|
|
|
self._new_pts_date = (None, None)
|
2018-02-15 13:41:32 +03:00
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# Some further state for subclasses
|
|
|
|
self._event_builders = []
|
2018-08-03 18:51:56 +03:00
|
|
|
self._conversations = {}
|
2018-10-12 23:17:07 +03:00
|
|
|
self._ids_in_conversations = {} # chat_id: count
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# Default parse mode
|
|
|
|
self._parse_mode = markdown
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# Some fields to easy signing in. Let {phone: hash} be
|
|
|
|
# a dictionary because the user may change their mind.
|
|
|
|
self._phone_code_hash = {}
|
|
|
|
self._phone = None
|
|
|
|
self._tos = None
|
2017-06-09 11:35:19 +03:00
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# Sometimes we need to know who we are, cache the self peer
|
|
|
|
self._self_input_peer = None
|
2018-12-24 19:32:34 +03:00
|
|
|
self._bot = None
|
2017-09-30 17:32:10 +03:00
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# endregion
|
2017-12-20 14:47:10 +03:00
|
|
|
|
2018-06-14 20:35:12 +03:00
|
|
|
# region Properties
|
|
|
|
|
|
|
|
@property
|
|
|
|
def loop(self):
|
|
|
|
return self._loop
|
|
|
|
|
2018-06-17 20:29:41 +03:00
|
|
|
@property
|
|
|
|
def disconnected(self):
|
|
|
|
"""
|
|
|
|
Future that resolves when the connection to Telegram
|
|
|
|
ends, either by user action or in the background.
|
|
|
|
"""
|
|
|
|
return self._sender.disconnected
|
|
|
|
|
2018-06-14 20:35:12 +03:00
|
|
|
# endregion
|
|
|
|
|
2018-06-08 22:52:59 +03:00
|
|
|
# region Connecting
|
2017-09-29 21:50:27 +03:00
|
|
|
|
2018-06-09 22:03:48 +03:00
|
|
|
async def connect(self):
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
|
|
|
Connects to Telegram.
|
|
|
|
"""
|
2018-10-19 14:50:11 +03:00
|
|
|
await self._sender.connect(self._connection(
|
2019-01-15 13:09:08 +03:00
|
|
|
self.session.server_address,
|
|
|
|
self.session.port,
|
2019-02-11 02:16:46 +03:00
|
|
|
self.session.dc_id,
|
2019-01-15 13:09:08 +03:00
|
|
|
loop=self._loop,
|
|
|
|
loggers=self._log,
|
2019-01-11 17:52:30 +03:00
|
|
|
proxy=self._proxy
|
2018-10-04 18:11:31 +03:00
|
|
|
))
|
2018-10-22 21:58:07 +03:00
|
|
|
self.session.auth_key = self._sender.auth_key
|
2018-10-24 17:39:59 +03:00
|
|
|
self.session.save()
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-06-17 17:23:22 +03:00
|
|
|
await self._sender.send(self._init_with(
|
|
|
|
functions.help.GetConfigRequest()))
|
|
|
|
|
2018-06-18 14:22:25 +03:00
|
|
|
self._updates_handle = self._loop.create_task(self._update_loop())
|
|
|
|
|
2017-09-17 17:39:29 +03:00
|
|
|
def is_connected(self):
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
|
|
|
Returns ``True`` if the user has connected.
|
|
|
|
"""
|
2018-06-28 17:04:12 +03:00
|
|
|
sender = getattr(self, '_sender', None)
|
|
|
|
return sender and sender.is_connected()
|
2017-09-17 17:39:29 +03:00
|
|
|
|
2018-10-16 12:56:17 +03:00
|
|
|
def disconnect(self):
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
|
|
|
Disconnects from Telegram.
|
2018-10-16 12:56:17 +03:00
|
|
|
|
2019-03-21 14:21:00 +03:00
|
|
|
If the event loop is already running, this method returns a
|
|
|
|
coroutine that you should await on your own code; otherwise
|
|
|
|
the loop is ran until said coroutine completes.
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
2019-03-21 14:21:00 +03:00
|
|
|
if self._loop.is_running():
|
|
|
|
return self._disconnect_coro()
|
|
|
|
else:
|
|
|
|
self._loop.run_until_complete(self._disconnect_coro())
|
|
|
|
|
|
|
|
async def _disconnect_coro(self):
|
|
|
|
await self._disconnect()
|
2019-03-28 14:32:02 +03:00
|
|
|
|
2019-04-10 20:09:15 +03:00
|
|
|
pts, date = self._new_pts_date
|
|
|
|
if pts:
|
|
|
|
self.session.set_update_state(0, types.updates.State(
|
|
|
|
pts=pts,
|
|
|
|
qts=0,
|
|
|
|
date=date or datetime.now(),
|
|
|
|
seq=0,
|
|
|
|
unread_count=0
|
|
|
|
))
|
2019-03-28 14:32:02 +03:00
|
|
|
|
2019-03-21 13:36:35 +03:00
|
|
|
self.session.close()
|
2018-06-26 14:37:34 +03:00
|
|
|
|
2019-03-21 14:21:00 +03:00
|
|
|
async def _disconnect(self):
|
2018-06-26 14:37:34 +03:00
|
|
|
"""
|
|
|
|
Disconnect only, without closing the session. Used in reconnections
|
|
|
|
to different data centers, where we don't want to close the session
|
|
|
|
file; user disconnects however should close it since it means that
|
|
|
|
their job with the client is complete and we should clean it up all.
|
|
|
|
"""
|
2019-03-21 14:21:00 +03:00
|
|
|
await self._sender.disconnect()
|
2019-04-01 09:46:07 +03:00
|
|
|
await helpers._cancel(self._log[__name__],
|
|
|
|
updates_handle=self._updates_handle)
|
2017-09-29 21:50:27 +03:00
|
|
|
|
2018-06-10 22:30:16 +03:00
|
|
|
async def _switch_dc(self, new_dc):
|
2017-06-08 17:51:20 +03:00
|
|
|
"""
|
2018-06-09 22:03:48 +03:00
|
|
|
Permanently switches the current connection to the new data center.
|
2018-06-08 22:52:59 +03:00
|
|
|
"""
|
2019-01-11 17:52:30 +03:00
|
|
|
self._log[__name__].info('Reconnecting to new data center %s', new_dc)
|
2018-06-10 22:30:16 +03:00
|
|
|
dc = await self._get_dc(new_dc)
|
2018-06-08 22:52:59 +03:00
|
|
|
|
2018-10-12 23:00:02 +03:00
|
|
|
self.session.set_dc(dc.id, dc.ip_address, dc.port)
|
2018-06-08 22:52:59 +03:00
|
|
|
# auth_key's are associated with a server, which has now changed
|
|
|
|
# so it's not valid anymore. Set to None to force recreating it.
|
2018-11-03 20:53:26 +03:00
|
|
|
self._sender.auth_key.key = None
|
2018-10-01 15:20:50 +03:00
|
|
|
self.session.auth_key = None
|
2018-10-12 23:00:02 +03:00
|
|
|
self.session.save()
|
2019-03-21 14:21:00 +03:00
|
|
|
await self._disconnect()
|
2018-06-10 22:30:16 +03:00
|
|
|
return await self.connect()
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-11-03 20:53:26 +03:00
|
|
|
def _auth_key_callback(self, auth_key):
|
2018-06-27 11:15:59 +03:00
|
|
|
"""
|
|
|
|
Callback from the sender whenever it needed to generate a
|
|
|
|
new authorization key. This means we are not authorized.
|
|
|
|
"""
|
|
|
|
self.session.auth_key = auth_key
|
2018-10-12 23:00:02 +03:00
|
|
|
self.session.save()
|
2018-06-27 11:15:59 +03:00
|
|
|
|
2017-06-08 14:12:57 +03:00
|
|
|
# endregion
|
|
|
|
|
2017-09-29 21:50:27 +03:00
|
|
|
# region Working with different connections/Data Centers
|
|
|
|
|
2018-06-09 22:03:48 +03:00
|
|
|
async def _get_dc(self, dc_id, cdn=False):
|
2017-06-08 14:12:57 +03:00
|
|
|
"""Gets the Data Center (DC) associated to 'dc_id'"""
|
2018-06-10 22:30:16 +03:00
|
|
|
cls = self.__class__
|
|
|
|
if not cls._config:
|
|
|
|
cls._config = await self(functions.help.GetConfigRequest())
|
|
|
|
|
|
|
|
if cdn and not self._cdn_config:
|
|
|
|
cls._cdn_config = await self(functions.help.GetCdnConfigRequest())
|
|
|
|
for pk in cls._cdn_config.public_keys:
|
|
|
|
rsa.add_key(pk.public_key)
|
|
|
|
|
|
|
|
return next(
|
|
|
|
dc for dc in cls._config.dc_options
|
|
|
|
if dc.id == dc_id
|
|
|
|
and bool(dc.ipv6) == self._use_ipv6 and bool(dc.cdn) == cdn
|
|
|
|
)
|
2017-06-08 14:12:57 +03:00
|
|
|
|
2018-06-28 15:10:36 +03:00
|
|
|
async def _create_exported_sender(self, dc_id):
|
2018-06-11 21:05:10 +03:00
|
|
|
"""
|
2018-06-28 15:10:36 +03:00
|
|
|
Creates a new exported `MTProtoSender` for the given `dc_id` and
|
|
|
|
returns it. This method should be used by `_borrow_exported_sender`.
|
2017-07-04 11:21:15 +03:00
|
|
|
"""
|
|
|
|
# Thanks badoualy/kotlogram on /telegram/api/DefaultTelegramClient.kt
|
2018-06-11 21:05:10 +03:00
|
|
|
# for clearly showing how to export the authorization
|
|
|
|
dc = await self._get_dc(dc_id)
|
|
|
|
# Can't reuse self._sender._connection as it has its own seqno.
|
|
|
|
#
|
|
|
|
# If one were to do that, Telegram would reset the connection
|
|
|
|
# with no further clues.
|
2019-01-12 14:15:29 +03:00
|
|
|
sender = MTProtoSender(None, self._loop, loggers=self._log)
|
2018-10-19 14:50:11 +03:00
|
|
|
await sender.connect(self._connection(
|
2019-01-15 13:09:08 +03:00
|
|
|
dc.ip_address,
|
|
|
|
dc.port,
|
2019-02-11 02:16:46 +03:00
|
|
|
dc.id,
|
2019-01-15 13:09:08 +03:00
|
|
|
loop=self._loop,
|
|
|
|
loggers=self._log,
|
|
|
|
proxy=self._proxy
|
|
|
|
))
|
2019-01-11 17:52:30 +03:00
|
|
|
self._log[__name__].info('Exporting authorization for data center %s',
|
|
|
|
dc)
|
2018-06-28 15:10:36 +03:00
|
|
|
auth = await self(functions.auth.ExportAuthorizationRequest(dc_id))
|
|
|
|
req = self._init_with(functions.auth.ImportAuthorizationRequest(
|
|
|
|
id=auth.id, bytes=auth.bytes
|
|
|
|
))
|
|
|
|
await sender.send(req)
|
|
|
|
return sender
|
|
|
|
|
|
|
|
async def _borrow_exported_sender(self, dc_id):
|
|
|
|
"""
|
|
|
|
Borrows a connected `MTProtoSender` for the given `dc_id`.
|
|
|
|
If it's not cached, creates a new one if it doesn't exist yet,
|
|
|
|
and imports a freshly exported authorization key for it to be usable.
|
|
|
|
|
|
|
|
Once its job is over it should be `_return_exported_sender`.
|
|
|
|
"""
|
|
|
|
async with self._borrow_sender_lock:
|
|
|
|
n, sender = self._borrowed_senders.get(dc_id, (0, None))
|
|
|
|
if not sender:
|
|
|
|
sender = await self._create_exported_sender(dc_id)
|
|
|
|
sender.dc_id = dc_id
|
2018-07-21 12:24:20 +03:00
|
|
|
elif not n:
|
|
|
|
dc = await self._get_dc(dc_id)
|
2018-10-20 18:11:40 +03:00
|
|
|
await sender.connect(self._connection(
|
2019-01-15 13:09:08 +03:00
|
|
|
dc.ip_address,
|
|
|
|
dc.port,
|
2019-02-11 02:16:46 +03:00
|
|
|
dc.id,
|
2019-01-15 13:09:08 +03:00
|
|
|
loop=self._loop,
|
|
|
|
loggers=self._log,
|
|
|
|
proxy=self._proxy
|
2018-10-20 18:11:40 +03:00
|
|
|
))
|
2018-06-28 15:10:36 +03:00
|
|
|
|
|
|
|
self._borrowed_senders[dc_id] = (n + 1, sender)
|
2017-10-24 16:40:51 +03:00
|
|
|
|
2018-06-11 21:05:10 +03:00
|
|
|
return sender
|
2017-07-04 11:21:15 +03:00
|
|
|
|
2018-06-28 15:10:36 +03:00
|
|
|
async def _return_exported_sender(self, sender):
|
|
|
|
"""
|
|
|
|
Returns a borrowed exported sender. If all borrows have
|
|
|
|
been returned, the sender is cleanly disconnected.
|
|
|
|
"""
|
|
|
|
async with self._borrow_sender_lock:
|
|
|
|
dc_id = sender.dc_id
|
|
|
|
n, _ = self._borrowed_senders[dc_id]
|
|
|
|
n -= 1
|
2018-07-21 12:24:20 +03:00
|
|
|
self._borrowed_senders[dc_id] = (n, sender)
|
|
|
|
if not n:
|
2019-01-11 17:52:30 +03:00
|
|
|
self._log[__name__].info(
|
|
|
|
'Disconnecting borrowed sender for DC %d', dc_id)
|
2019-03-21 14:21:00 +03:00
|
|
|
await sender.disconnect()
|
2018-06-28 15:10:36 +03:00
|
|
|
|
2018-06-09 22:03:48 +03:00
|
|
|
async def _get_cdn_client(self, cdn_redirect):
|
2018-06-28 15:10:36 +03:00
|
|
|
"""Similar to ._borrow_exported_client, but for CDNs"""
|
2018-06-09 22:03:48 +03:00
|
|
|
# TODO Implement
|
|
|
|
raise NotImplementedError
|
2017-09-30 18:51:07 +03:00
|
|
|
session = self._exported_sessions.get(cdn_redirect.dc_id)
|
|
|
|
if not session:
|
2018-06-09 22:03:48 +03:00
|
|
|
dc = await self._get_dc(cdn_redirect.dc_id, cdn=True)
|
2018-10-12 23:00:02 +03:00
|
|
|
session = self.session.clone()
|
2018-10-05 21:25:49 +03:00
|
|
|
await session.set_dc(dc.id, dc.ip_address, dc.port)
|
2017-09-30 18:51:07 +03:00
|
|
|
self._exported_sessions[cdn_redirect.dc_id] = session
|
|
|
|
|
2019-01-12 14:15:29 +03:00
|
|
|
self._log[__name__].info('Creating new CDN client')
|
2019-03-12 03:28:59 +03:00
|
|
|
client = TelegramBareClient(
|
2017-09-30 18:51:07 +03:00
|
|
|
session, self.api_id, self.api_hash,
|
|
|
|
proxy=self._sender.connection.conn.proxy,
|
|
|
|
timeout=self._sender.connection.get_timeout()
|
|
|
|
)
|
|
|
|
|
|
|
|
# This will make use of the new RSA keys for this specific CDN.
|
|
|
|
#
|
2017-10-24 16:40:51 +03:00
|
|
|
# We won't be calling GetConfigRequest because it's only called
|
|
|
|
# when needed by ._get_dc, and also it's static so it's likely
|
|
|
|
# set already. Avoid invoking non-CDN methods by not syncing updates.
|
|
|
|
client.connect(_sync_updates=False)
|
2017-09-30 18:51:07 +03:00
|
|
|
return client
|
|
|
|
|
2017-06-08 14:12:57 +03:00
|
|
|
# endregion
|
|
|
|
|
|
|
|
# region Invoking Telegram requests
|
|
|
|
|
2018-06-09 22:22:54 +03:00
|
|
|
@abc.abstractmethod
|
2018-06-18 19:11:16 +03:00
|
|
|
def __call__(self, request, ordered=False):
|
2018-05-09 11:19:45 +03:00
|
|
|
"""
|
|
|
|
Invokes (sends) one or more MTProtoRequests and returns (receives)
|
|
|
|
their result.
|
|
|
|
|
|
|
|
Args:
|
|
|
|
request (`TLObject` | `list`):
|
|
|
|
The request or requests to be invoked.
|
|
|
|
|
|
|
|
ordered (`bool`, optional):
|
|
|
|
Whether the requests (if more than one was given) should be
|
|
|
|
executed sequentially on the server. They run in arbitrary
|
|
|
|
order by default.
|
|
|
|
|
|
|
|
Returns:
|
|
|
|
The result of the request (often a `TLObject`) or a list of
|
|
|
|
results if more than one request was given.
|
2017-06-08 14:12:57 +03:00
|
|
|
"""
|
2018-06-09 22:22:54 +03:00
|
|
|
raise NotImplementedError
|
2017-09-30 12:45:35 +03:00
|
|
|
|
2018-06-13 17:20:15 +03:00
|
|
|
@abc.abstractmethod
|
|
|
|
def _handle_update(self, update):
|
|
|
|
raise NotImplementedError
|
|
|
|
|
2018-06-18 14:22:25 +03:00
|
|
|
@abc.abstractmethod
|
|
|
|
def _update_loop(self):
|
|
|
|
raise NotImplementedError
|
|
|
|
|
2018-06-27 20:40:32 +03:00
|
|
|
@abc.abstractmethod
|
|
|
|
async def _handle_auto_reconnect(self):
|
|
|
|
raise NotImplementedError
|
|
|
|
|
2017-09-29 21:50:27 +03:00
|
|
|
# endregion
|