Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 4 additions & 4 deletions core/db_adapter/aioredis_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

from core.db_adapter.db_adapter import AsyncDBAdapter
from core.db_adapter import error
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring

from core.logging.logger_utils import log

Expand All @@ -21,15 +21,15 @@ def __init__(self, config=None):
except KeyError:
pass

@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
async def save(self, id, data):
return await self._async_run(self._save, id, data)

@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
async def replace_if_equals(self, id, sample, data):
return await self._async_run(self._replace_if_equals, id, sample, data)

@monitoring.got_histogram_decorate("get_time")
@monitoring.monitoring.got_histogram_decorate("get_time")
async def get(self, id):
return await self._async_run(self._get, id)

Expand Down
8 changes: 4 additions & 4 deletions core/db_adapter/aioredis_sentinel_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
from aioredis.sentinel import Sentinel
from core.db_adapter.db_adapter import AsyncDBAdapter
from core.db_adapter import error
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.logging.logger_utils import log


Expand All @@ -22,16 +22,16 @@ def __init__(self, config=None):
except KeyError:
pass

@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
async def save(self, id, data):
return await self._async_run(self._save, id, data)


@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
async def replace_if_equals(self, id, sample, data):
return await self._async_run(self._replace_if_equals, id, sample, data)

@monitoring.got_histogram_decorate("get_time")
@monitoring.monitoring.got_histogram_decorate("get_time")
async def get(self, id):
return await self._async_run(self._get, id)

Expand Down
4 changes: 2 additions & 2 deletions core/db_adapter/ceph/ceph_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
from core.db_adapter.ceph.ceph_io import CephIO
from core.db_adapter.db_adapter import DBAdapter
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring

ssl._create_default_https_context = ssl._create_unverified_context

Expand All @@ -35,7 +35,7 @@ def connect(self):
params={log_const.KEY_NAME: log_const.HANDLED_EXCEPTION_VALUE},
level="ERROR",
exc_info=True)
monitoring.got_counter("ceph_connection_exception")
monitoring.monitoring.got_counter("ceph_connection_exception")
raise

@property
Expand Down
16 changes: 8 additions & 8 deletions core/db_adapter/db_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
from core.logging.logger_utils import log
from core.model.factory import build_factory
from core.model.registered import Registered
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.utils.rerunable import Rerunable

db_adapters = Registered()
Expand Down Expand Up @@ -65,15 +65,15 @@ def path_exists(self, path):
def mtime(self, path):
return self._run(self._mtime, path)

@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
def save(self, id, data):
return self._run(self._save, id, data)

@monitoring.got_histogram_decorate("save_time")
@monitoring.monitoring.got_histogram_decorate("save_time")
def replace_if_equals(self, id, sample, data):
return self._run(self._replace_if_equals, id, sample, data)

@monitoring.got_histogram_decorate("get_time")
@monitoring.monitoring.got_histogram_decorate("get_time")
def get(self, id):
return self._run(self._get, id)

Expand Down Expand Up @@ -109,15 +109,15 @@ async def _path_exists(self, path):
async def path_exists(self, path):
return await self._async_run(self._path_exists, path)

@monitoring.got_histogram("save_time")
@monitoring.monitoring.got_histogram("save_time")
async def save(self, id, data):
return await self._async_run(self._save, id, data)

@monitoring.got_histogram("save_time")
@monitoring.monitoring.got_histogram("save_time")
async def replace_if_equals(self, id, sample, data):
return await self._async_run(self._replace_if_equals, id, sample, data)

@monitoring.got_histogram("get_time")
@monitoring.monitoring.got_histogram("get_time")
async def get(self, id):
return await self._async_run(self._get, id)

Expand All @@ -144,5 +144,5 @@ async def _async_run(self, action, *args, _try_count=None, **kwargs):
result = await self._async_run(action, *args, _try_count=_try_count, **kwargs)
counter_name = self._get_counter_name()
if counter_name:
monitoring.got_counter(f"{counter_name}_exception")
monitoring.monitoring.got_counter(f"{counter_name}_exception")
return result
6 changes: 3 additions & 3 deletions core/db_adapter/ignite_adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
from core.db_adapter import error
from core.db_adapter.db_adapter import AsyncDBAdapter
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring


class IgniteAdapter(AsyncDBAdapter):
Expand Down Expand Up @@ -55,7 +55,7 @@ async def connect(self):
params={log_const.KEY_NAME: log_const.HANDLED_EXCEPTION_VALUE},
level="ERROR",
exc_info=True)
monitoring.got_counter("ignite_connection_exception")
monitoring.monitoring.got_counter("ignite_connection_exception")
raise

async def _save(self, id, data):
Expand All @@ -75,7 +75,7 @@ async def get_cache(self):
if self._client is None:
log('Attempt to recreate ignite instance', level="WARNING")
await self.connect()
monitoring.got_counter("ignite_reconnection")
monitoring.monitoring.got_counter("ignite_reconnection")
return self._cache

@property
Expand Down
6 changes: 3 additions & 3 deletions core/mq/kafka/async_kafka_publisher.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@

import core.logging.logger_constants as log_const
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.mq.kafka.kafka_publisher import KafkaPublisher


Expand All @@ -30,7 +30,7 @@ def send(self, value, key=None, topic_key=None, headers=None):
}
log("KafkaProducer: Local producer queue is full (%(queue_amount)s messages awaiting delivery):"
" try again\n", params=params, level="ERROR")
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")

def send_to_topic(self, value, key=None, topic=None, headers=None):
try:
Expand All @@ -52,7 +52,7 @@ def send_to_topic(self, value, key=None, topic=None, headers=None):
}
log("KafkaProducer: Local producer queue is full (%(queue_amount)s messages awaiting delivery):"
" try again\n", params=params, level="ERROR")
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")

def _poll_for_callbacks(self):
poll_timeout = self._config.get("poll_timeout", 1)
Expand Down
6 changes: 3 additions & 3 deletions core/mq/kafka/kafka_consumer.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@

import core.logging.logger_constants as log_const
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.mq.kafka.base_kafka_consumer import BaseKafkaConsumer


Expand Down Expand Up @@ -102,7 +102,7 @@ def _error_callback(self, err):
log_const.KEY_NAME: log_const.EXCEPTION_VALUE
}
log("KafkaConsumer: Error: %(error)s", params=params, level="WARNING")
monitoring.got_counter("kafka_consumer_exception")
monitoring.monitoring.got_counter("kafka_consumer_exception")

# noinspection PyMethodMayBeStatic
def _process_message(self, msg: KafkaMessage):
Expand All @@ -111,7 +111,7 @@ def _process_message(self, msg: KafkaMessage):
if err.code() == KafkaError._PARTITION_EOF:
return None
else:
monitoring.got_counter("kafka_consumer_exception")
monitoring.monitoring.got_counter("kafka_consumer_exception")
params = {
"code": err.code(),
"pid": os.getpid(),
Expand Down
10 changes: 5 additions & 5 deletions core/mq/kafka/kafka_publisher.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@

import core.logging.logger_constants as log_const
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.mq.kafka.base_kafka_publisher import BaseKafkaPublisher


Expand Down Expand Up @@ -41,7 +41,7 @@ def send(self, value, key=None, topic_key=None, headers=None):
}
log("KafkaProducer: Local producer queue is full (%(queue_amount)s messages awaiting delivery):"
" try again\n", params=params, level="ERROR")
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")
self._poll()

def send_to_topic(self, value, key=None, topic=None, headers=None):
Expand All @@ -64,7 +64,7 @@ def send_to_topic(self, value, key=None, topic=None, headers=None):
}
log("KafkaProducer: Local producer queue is full (%(queue_amount)s messages awaiting delivery):"
" try again\n", params=params, level="ERROR")
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")
self._poll()

def _poll(self):
Expand All @@ -81,7 +81,7 @@ def _error_callback(self, err):
log_const.KEY_NAME: log_const.EXCEPTION_VALUE
}
log("KafkaProducer: Error: %(error)s", params=params, level="ERROR")
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")

def _delivery_callback(self, err, msg):
if err:
Expand All @@ -101,7 +101,7 @@ def _delivery_callback(self, err, msg):
log_const.KEY_NAME: log_const.EXCEPTION_VALUE},
level="ERROR",
exc_info=True)
monitoring.got_counter("kafka_producer_exception")
monitoring.monitoring.got_counter("kafka_producer_exception")

def close(self):
self._producer.flush(self._config["flush_timeout"])
4 changes: 2 additions & 2 deletions core/request/rest_request.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
from timeout_decorator import timeout_decorator
from core.request.base_request import BaseRequest
from core.utils.exception_handlers import exc_handler
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring


class RestNoMethodSpecifiedException(Exception):
Expand Down Expand Up @@ -37,7 +37,7 @@ def run(self, data, params=None):
return method(data)

def on_timeout_error(self, *args, **kwarg):
monitoring.got_counter("core_rest_run_timeout")
monitoring.monitoring.got_counter("core_rest_run_timeout")

def _requests_get(self, params):
return requests.get(self.url, params=params, **self.rest_args).text
Expand Down
4 changes: 2 additions & 2 deletions core/unified_template/unified_template.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

import core.logging.logger_constants as log_const
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring

UNIFIED_TEMPLATE_TYPE_NAME = "unified_template"

Expand Down Expand Up @@ -49,7 +49,7 @@ def render(self, *args, **kwargs):
"params_dict_str": str(params_dict)},
level="ERROR",
exc_info=True)
monitoring.got_counter("core_jinja_template_error")
monitoring.monitoring.got_counter("core_jinja_template_error")
raise
return result

Expand Down
4 changes: 2 additions & 2 deletions core/utils/rerunable.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@

import core.logging.logger_constants as log_const
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring


class Rerunable():
Expand Down Expand Up @@ -44,7 +44,7 @@ def _run(self, action, *args, _try_count=None, **kwargs):
result = self._run(action, *args, _try_count=_try_count, **kwargs)
counter_name = self._get_counter_name()
if counter_name:
monitoring.got_counter(f"{counter_name}_exception")
monitoring.monitoring.got_counter(f"{counter_name}_exception")
return result

def _get_counter_name(self):
Expand Down
4 changes: 2 additions & 2 deletions scenarios/scenario_descriptions/form_filling_scenario.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
from typing import Dict, Any

from core.basic_models.scenarios.base_scenario import BaseScenario
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.logging.logger_utils import log

import scenarios.logging.logger_constants as log_const
Expand Down Expand Up @@ -174,7 +174,7 @@ async def get_reply(self, user, text_preprocessing_result, reply_actions, field,
user.last_scenarios.delete(self.id)
return action_messages

@monitoring.got_histogram_decorate("scenario_time")
@monitoring.monitoring.got_histogram_decorate("scenario_time")
async def run(self, text_preprocessing_result, user, params: Dict[str, Any] = None):
form = self._get_form(user)
user.last_scenarios.add(self.id, text_preprocessing_result)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@
from scenarios.scenario_descriptions.form_filling_scenario import FormFillingScenario
from scenarios.scenario_descriptions.tree_scenario.tree_scenario_node import TreeScenarioNode
from core.model.factory import dict_factory
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.logging.logger_utils import log
import scenarios.logging.logger_constants as log_const
from scenarios.scenario_models.history import Event, HistoryConstants
Expand Down Expand Up @@ -83,7 +83,7 @@ def get_fields_data(self, main_form, form_type):
all_forms_fields.update(form_field_data)
return all_forms_fields

@monitoring.got_histogram_decorate("scenario_time")
@monitoring.monitoring.got_histogram_decorate("scenario_time")
async def run(self, text_preprocessing_result, user, params: Dict[str, Any] = None):
main_form = self._get_form(user)
user.last_scenarios.add(self.id, text_preprocessing_result)
Expand Down
5 changes: 2 additions & 3 deletions smart_kit/start_points/base_main_loop.py
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
from core.db_adapter.db_adapter import DBAdapterException
from core.db_adapter.db_adapter import db_adapter_factory
from core.logging.logger_utils import log
from core.monitoring.monitoring import monitoring
from core.monitoring import monitoring
from core.monitoring.healthcheck_handler import RootResource
from core.monitoring.twisted_server import TwistedServer
from core.model.base_user import BaseUser
Expand Down Expand Up @@ -95,8 +95,7 @@ def _create_health_check_server(self, settings):

def _init_monitoring_config(self, template_settings):
monitoring_config = template_settings["monitoring"]
monitoring.apply_config(monitoring_config)
smart_kit_metrics.apply_config(monitoring_config)
monitoring.monitoring.apply_config(monitoring_config)
smart_kit_metrics.init_metrics(app_name=self.app_name)

async def load_user(self, db_uid, message):
Expand Down
Loading