
The change updates the imports to use simplejson library and fixes three unit tests. In particular TestUtils.test_parse_alarm_state_hist_message has been slightly changed to make testing more robust. Also, Python 2 specific implemetation has been removed from influxdb.line_utils . Additionally, standard library unittest.mock is used instead of the third party mock lib. Depends-On: https://review.opendev.org/718014 Depends-On: https://review.opendev.org/720188 Change-Id: I64b1a60e8be929c25c005a0f429b1274cb8570e6 Story: 2007549 Task: 39390
78 lines
2.7 KiB
Python
78 lines
2.7 KiB
Python
# Copyright 2017 FUJITSU LIMITED
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
# implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
from datetime import datetime
|
|
|
|
from elasticsearch import Elasticsearch
|
|
from oslo_config import cfg
|
|
from oslo_log import log
|
|
import simplejson as json
|
|
|
|
from monasca_persister.repositories import abstract_repository
|
|
from monasca_persister.repositories import data_points
|
|
from monasca_persister.repositories import utils
|
|
|
|
LOG = log.getLogger(__name__)
|
|
|
|
|
|
class ElasticSearchEventsRepository(abstract_repository.AbstractRepository):
|
|
def __init__(self):
|
|
super(ElasticSearchEventsRepository, self).__init__()
|
|
self.conf = cfg.CONF.elasticsearch
|
|
self.es = Elasticsearch(
|
|
hosts=self.conf.hosts,
|
|
sniff_on_start=self.conf.sniff_on_start,
|
|
sniff_on_connection_fail=self.conf.sniff_on_connection_fail,
|
|
sniffer_timeout=self.conf.sniffer_timeout,
|
|
max_retries=self.conf.max_retries
|
|
)
|
|
self.data_points_class = data_points.DataPointsAsList
|
|
|
|
def process_message(self, message):
|
|
return utils.parse_events_message(message)
|
|
|
|
def write_batch(self, data_points):
|
|
for data_point in data_points:
|
|
(project_id, timestamp, event_type, payload, dimensions) = data_point
|
|
|
|
index = '%s-%s-%s' % (self.conf.index_name, project_id,
|
|
ElasticSearchEventsRepository._normalize_timestamp(timestamp))
|
|
|
|
body = {
|
|
'project_id': project_id,
|
|
'@timestamp': timestamp,
|
|
'event_type': event_type,
|
|
'payload': payload,
|
|
'dimensions': dimensions
|
|
}
|
|
|
|
self.es.create(
|
|
index=index,
|
|
doc_type='event',
|
|
body=json.dumps(body)
|
|
)
|
|
|
|
@staticmethod
|
|
def _normalize_timestamp(timestamp):
|
|
d = None
|
|
if timestamp and len(timestamp) >= 10:
|
|
try:
|
|
d = datetime.strptime(timestamp[0:10], '%Y-%m-%d')
|
|
except ValueError as e:
|
|
LOG.warning("Unable to parse timestamp '%s' - %s" % (timestamp, str(e)))
|
|
if not d:
|
|
d = datetime.today()
|
|
return d.strftime('%Y-%m-%d')
|