Datasets:

Modalities:
Text
Formats:
json
Languages:
code
Size:
< 1K
Tags:
code
Libraries:
Datasets
pandas
License:
File size: 16,725 Bytes
4fb7874
1
# -*- coding: utf-8 -*-# Copyright 2015, 2016 OpenMarket Ltd# Copyright 2017 New Vector Ltd## Licensed under the Apache License, Version 2.0 (the "License");# you may not use this file except in compliance with the License.# You may obtain a copy of the License at##     http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing, software# distributed under the License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.# See the License for the specific language governing permissions and# limitations under the License.import loggingfrom prometheus_client import Counterfrom twisted.internet.error import AlreadyCalled, AlreadyCancelledfrom synapse.api.constants import EventTypesfrom synapse.logging import opentracingfrom synapse.metrics.background_process_metrics import run_as_background_processfrom synapse.push import PusherConfigExceptionfrom synapse.types import RoomStreamTokenfrom . import push_rule_evaluator, push_toolslogger = logging.getLogger(__name__)http_push_processed_counter = Counter(    "synapse_http_httppusher_http_pushes_processed",    "Number of push notifications successfully sent",)http_push_failed_counter = Counter(    "synapse_http_httppusher_http_pushes_failed",    "Number of push notifications which failed",)http_badges_processed_counter = Counter(    "synapse_http_httppusher_badge_updates_processed",    "Number of badge updates successfully sent",)http_badges_failed_counter = Counter(    "synapse_http_httppusher_badge_updates_failed",    "Number of badge updates which failed",)class HttpPusher:    INITIAL_BACKOFF_SEC = 1  # in seconds because that's what Twisted takes    MAX_BACKOFF_SEC = 60 * 60    # This one's in ms because we compare it against the clock    GIVE_UP_AFTER_MS = 24 * 60 * 60 * 1000    def __init__(self, hs, pusherdict):        self.hs = hs        self.store = self.hs.get_datastore()        self.storage = self.hs.get_storage()        self.clock = self.hs.get_clock()        self.state_handler = self.hs.get_state_handler()        self.user_id = pusherdict["user_name"]        self.app_id = pusherdict["app_id"]        self.app_display_name = pusherdict["app_display_name"]        self.device_display_name = pusherdict["device_display_name"]        self.pushkey = pusherdict["pushkey"]        self.pushkey_ts = pusherdict["ts"]        self.data = pusherdict["data"]        self.last_stream_ordering = pusherdict["last_stream_ordering"]        self.backoff_delay = HttpPusher.INITIAL_BACKOFF_SEC        self.failing_since = pusherdict["failing_since"]        self.timed_call = None        self._is_processing = False        self._group_unread_count_by_room = hs.config.push_group_unread_count_by_room        # This is the highest stream ordering we know it's safe to process.        # When new events arrive, we'll be given a window of new events: we        # should honour this rather than just looking for anything higher        # because of potential out-of-order event serialisation. This starts        # off as None though as we don't know any better.        self.max_stream_ordering = None        if "data" not in pusherdict:            raise PusherConfigException("No 'data' key for HTTP pusher")        self.data = pusherdict["data"]        self.name = "%s/%s/%s" % (            pusherdict["user_name"],            pusherdict["app_id"],            pusherdict["pushkey"],        )        if self.data is None:            raise PusherConfigException("data can not be null for HTTP pusher")        if "url" not in self.data:            raise PusherConfigException("'url' required in data for HTTP pusher")        self.url = self.data["url"]        # BUG: CWE-601 URL Redirection to Untrusted Site ('Open Redirect')        # self.http_client = hs.get_proxied_http_client()        # FIXED:         self.http_client = hs.get_proxied_blacklisted_http_client()        self.data_minus_url = {}        self.data_minus_url.update(self.data)        del self.data_minus_url["url"]    def on_started(self, should_check_for_notifs):        """Called when this pusher has been started.        Args:            should_check_for_notifs (bool): Whether we should immediately                check for push to send. Set to False only if it's known there                is nothing to send        """        if should_check_for_notifs:            self._start_processing()    def on_new_notifications(self, max_token: RoomStreamToken):        # We just use the minimum stream ordering and ignore the vector clock        # component. This is safe to do as long as we *always* ignore the vector        # clock components.        max_stream_ordering = max_token.stream        self.max_stream_ordering = max(            max_stream_ordering, self.max_stream_ordering or 0        )        self._start_processing()    def on_new_receipts(self, min_stream_id, max_stream_id):        # Note that the min here shouldn't be relied upon to be accurate.        # We could check the receipts are actually m.read receipts here,        # but currently that's the only type of receipt anyway...        run_as_background_process("http_pusher.on_new_receipts", self._update_badge)    async def _update_badge(self):        # XXX as per https://github.com/matrix-org/matrix-doc/issues/2627, this seems        # to be largely redundant. perhaps we can remove it.        badge = await push_tools.get_badge_count(            self.hs.get_datastore(),            self.user_id,            group_by_room=self._group_unread_count_by_room,        )        await self._send_badge(badge)    def on_timer(self):        self._start_processing()    def on_stop(self):        if self.timed_call:            try:                self.timed_call.cancel()            except (AlreadyCalled, AlreadyCancelled):                pass            self.timed_call = None    def _start_processing(self):        if self._is_processing:            return        run_as_background_process("httppush.process", self._process)    async def _process(self):        # we should never get here if we are already processing        assert not self._is_processing        try:            self._is_processing = True            # if the max ordering changes while we're running _unsafe_process,            # call it again, and so on until we've caught up.            while True:                starting_max_ordering = self.max_stream_ordering                try:                    await self._unsafe_process()                except Exception:                    logger.exception("Exception processing notifs")                if self.max_stream_ordering == starting_max_ordering:                    break        finally:            self._is_processing = False    async def _unsafe_process(self):        """        Looks for unset notifications and dispatch them, in order        Never call this directly: use _process which will only allow this to        run once per pusher.        """        fn = self.store.get_unread_push_actions_for_user_in_range_for_http        unprocessed = await fn(            self.user_id, self.last_stream_ordering, self.max_stream_ordering        )        logger.info(            "Processing %i unprocessed push actions for %s starting at "            "stream_ordering %s",            len(unprocessed),            self.name,            self.last_stream_ordering,        )        for push_action in unprocessed:            with opentracing.start_active_span(                "http-push",                tags={                    "authenticated_entity": self.user_id,                    "event_id": push_action["event_id"],                    "app_id": self.app_id,                    "app_display_name": self.app_display_name,                },            ):                processed = await self._process_one(push_action)            if processed:                http_push_processed_counter.inc()                self.backoff_delay = HttpPusher.INITIAL_BACKOFF_SEC                self.last_stream_ordering = push_action["stream_ordering"]                pusher_still_exists = await self.store.update_pusher_last_stream_ordering_and_success(                    self.app_id,                    self.pushkey,                    self.user_id,                    self.last_stream_ordering,                    self.clock.time_msec(),                )                if not pusher_still_exists:                    # The pusher has been deleted while we were processing, so                    # lets just stop and return.                    self.on_stop()                    return                if self.failing_since:                    self.failing_since = None                    await self.store.update_pusher_failing_since(                        self.app_id, self.pushkey, self.user_id, self.failing_since                    )            else:                http_push_failed_counter.inc()                if not self.failing_since:                    self.failing_since = self.clock.time_msec()                    await self.store.update_pusher_failing_since(                        self.app_id, self.pushkey, self.user_id, self.failing_since                    )                if (                    self.failing_since                    and self.failing_since                    < self.clock.time_msec() - HttpPusher.GIVE_UP_AFTER_MS                ):                    # we really only give up so that if the URL gets                    # fixed, we don't suddenly deliver a load                    # of old notifications.                    logger.warning(                        "Giving up on a notification to user %s, pushkey %s",                        self.user_id,                        self.pushkey,                    )                    self.backoff_delay = HttpPusher.INITIAL_BACKOFF_SEC                    self.last_stream_ordering = push_action["stream_ordering"]                    pusher_still_exists = await self.store.update_pusher_last_stream_ordering(                        self.app_id,                        self.pushkey,                        self.user_id,                        self.last_stream_ordering,                    )                    if not pusher_still_exists:                        # The pusher has been deleted while we were processing, so                        # lets just stop and return.                        self.on_stop()                        return                    self.failing_since = None                    await self.store.update_pusher_failing_since(                        self.app_id, self.pushkey, self.user_id, self.failing_since                    )                else:                    logger.info("Push failed: delaying for %ds", self.backoff_delay)                    self.timed_call = self.hs.get_reactor().callLater(                        self.backoff_delay, self.on_timer                    )                    self.backoff_delay = min(                        self.backoff_delay * 2, self.MAX_BACKOFF_SEC                    )                    break    async def _process_one(self, push_action):        if "notify" not in push_action["actions"]:            return True        tweaks = push_rule_evaluator.tweaks_for_actions(push_action["actions"])        badge = await push_tools.get_badge_count(            self.hs.get_datastore(),            self.user_id,            group_by_room=self._group_unread_count_by_room,        )        event = await self.store.get_event(push_action["event_id"], allow_none=True)        if event is None:            return True  # It's been redacted        rejected = await self.dispatch_push(event, tweaks, badge)        if rejected is False:            return False        if isinstance(rejected, list) or isinstance(rejected, tuple):            for pk in rejected:                if pk != self.pushkey:                    # for sanity, we only remove the pushkey if it                    # was the one we actually sent...                    logger.warning(                        ("Ignoring rejected pushkey %s because we didn't send it"), pk,                    )                else:                    logger.info("Pushkey %s was rejected: removing", pk)                    await self.hs.remove_pusher(self.app_id, pk, self.user_id)        return True    async def _build_notification_dict(self, event, tweaks, badge):        priority = "low"        if (            event.type == EventTypes.Encrypted            or tweaks.get("highlight")            or tweaks.get("sound")        ):            # HACK send our push as high priority only if it generates a sound, highlight            #  or may do so (i.e. is encrypted so has unknown effects).            priority = "high"        if self.data.get("format") == "event_id_only":            d = {                "notification": {                    "event_id": event.event_id,                    "room_id": event.room_id,                    "counts": {"unread": badge},                    "prio": priority,                    "devices": [                        {                            "app_id": self.app_id,                            "pushkey": self.pushkey,                            "pushkey_ts": int(self.pushkey_ts / 1000),                            "data": self.data_minus_url,                        }                    ],                }            }            return d        ctx = await push_tools.get_context_for_event(            self.storage, self.state_handler, event, self.user_id        )        d = {            "notification": {                "id": event.event_id,  # deprecated: remove soon                "event_id": event.event_id,                "room_id": event.room_id,                "type": event.type,                "sender": event.user_id,                "prio": priority,                "counts": {                    "unread": badge,                    # 'missed_calls': 2                },                "devices": [                    {                        "app_id": self.app_id,                        "pushkey": self.pushkey,                        "pushkey_ts": int(self.pushkey_ts / 1000),                        "data": self.data_minus_url,                        "tweaks": tweaks,                    }                ],            }        }        if event.type == "m.room.member" and event.is_state():            d["notification"]["membership"] = event.content["membership"]            d["notification"]["user_is_target"] = event.state_key == self.user_id        if self.hs.config.push_include_content and event.content:            d["notification"]["content"] = event.content        # We no longer send aliases separately, instead, we send the human        # readable name of the room, which may be an alias.        if "sender_display_name" in ctx and len(ctx["sender_display_name"]) > 0:            d["notification"]["sender_display_name"] = ctx["sender_display_name"]        if "name" in ctx and len(ctx["name"]) > 0:            d["notification"]["room_name"] = ctx["name"]        return d    async def dispatch_push(self, event, tweaks, badge):        notification_dict = await self._build_notification_dict(event, tweaks, badge)        if not notification_dict:            return []        try:            resp = await self.http_client.post_json_get_json(                self.url, notification_dict            )        except Exception as e:            logger.warning(                "Failed to push event %s to %s: %s %s",                event.event_id,                self.name,                type(e),                e,            )            return False        rejected = []        if "rejected" in resp:            rejected = resp["rejected"]        return rejected    async def _send_badge(self, badge):        """        Args:            badge (int): number of unread messages        """        logger.debug("Sending updated badge count %d to %s", badge, self.name)        d = {            "notification": {                "id": "",                "type": None,                "sender": "",                "counts": {"unread": badge},                "devices": [                    {                        "app_id": self.app_id,                        "pushkey": self.pushkey,                        "pushkey_ts": int(self.pushkey_ts / 1000),                        "data": self.data_minus_url,                    }                ],            }        }        try:            await self.http_client.post_json_get_json(self.url, d)            http_badges_processed_counter.inc()        except Exception as e:            logger.warning(                "Failed to send badge count to %s: %s %s", self.name, type(e), e            )            http_badges_failed_counter.inc()