2022-04-04 16:50:17 +02:00
|
|
|
from logging import getLogger
|
|
|
|
from queue import Queue, Empty
|
2022-04-11 00:01:21 +02:00
|
|
|
from threading import Thread, Event, RLock
|
2022-04-04 16:50:17 +02:00
|
|
|
from time import time
|
2022-04-11 00:01:21 +02:00
|
|
|
from typing import Iterable, List, Optional
|
2022-04-04 16:50:17 +02:00
|
|
|
|
2022-04-05 23:04:19 +02:00
|
|
|
from sqlalchemy import and_, or_
|
2022-04-11 00:01:21 +02:00
|
|
|
from sqlalchemy.orm import Session, make_transient
|
2022-04-04 16:50:17 +02:00
|
|
|
|
2022-04-07 01:46:37 +02:00
|
|
|
from platypush.context import get_bus
|
|
|
|
from platypush.message.event.entities import EntityUpdateEvent
|
|
|
|
|
2022-04-04 16:50:17 +02:00
|
|
|
from ._base import Entity
|
|
|
|
|
|
|
|
|
|
|
|
class EntitiesEngine(Thread):
|
|
|
|
# Processing queue timeout in seconds
|
2022-04-05 21:17:58 +02:00
|
|
|
_queue_timeout = 5.0
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
obj_name = self.__class__.__name__
|
|
|
|
super().__init__(name=obj_name)
|
|
|
|
self.logger = getLogger(name=obj_name)
|
|
|
|
self._queue = Queue()
|
|
|
|
self._should_stop = Event()
|
2022-04-11 00:01:21 +02:00
|
|
|
self._entities_cache_lock = RLock()
|
|
|
|
self._entities_cache = {
|
|
|
|
'by_id': {},
|
|
|
|
'by_external_id_and_plugin': {},
|
|
|
|
'by_name_and_plugin': {},
|
|
|
|
}
|
|
|
|
|
|
|
|
def _get_db(self):
|
|
|
|
from platypush.context import get_plugin
|
|
|
|
|
|
|
|
db = get_plugin('db')
|
|
|
|
assert db
|
|
|
|
return db
|
|
|
|
|
|
|
|
def _get_cached_entity(self, entity: Entity) -> Optional[dict]:
|
|
|
|
if entity.id:
|
|
|
|
e = self._entities_cache['by_id'].get(entity.id)
|
|
|
|
if e:
|
|
|
|
return e
|
|
|
|
|
|
|
|
if entity.external_id and entity.plugin:
|
|
|
|
e = self._entities_cache['by_external_id_and_plugin'].get(
|
|
|
|
(entity.external_id, entity.plugin)
|
|
|
|
)
|
|
|
|
if e:
|
|
|
|
return e
|
|
|
|
|
|
|
|
if entity.name and entity.plugin:
|
|
|
|
e = self._entities_cache['by_name_and_plugin'].get(
|
|
|
|
(entity.name, entity.plugin)
|
|
|
|
)
|
|
|
|
if e:
|
|
|
|
return e
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def _cache_repr(entity: Entity) -> dict:
|
|
|
|
repr_ = entity.to_json()
|
|
|
|
repr_.pop('data', None)
|
|
|
|
repr_.pop('meta', None)
|
|
|
|
repr_.pop('created_at', None)
|
|
|
|
repr_.pop('updated_at', None)
|
|
|
|
return repr_
|
|
|
|
|
|
|
|
def _cache_entities(self, *entities: Entity, overwrite_cache=False):
|
|
|
|
for entity in entities:
|
|
|
|
e = self._cache_repr(entity)
|
|
|
|
if not overwrite_cache:
|
|
|
|
existing_entity = self._entities_cache['by_id'].get(entity.id)
|
|
|
|
if existing_entity:
|
|
|
|
for k, v in existing_entity.items():
|
|
|
|
if e.get(k) is None:
|
|
|
|
e[k] = v
|
|
|
|
|
|
|
|
if entity.id:
|
|
|
|
self._entities_cache['by_id'][entity.id] = e
|
|
|
|
if entity.external_id and entity.plugin:
|
|
|
|
self._entities_cache['by_external_id_and_plugin'][
|
|
|
|
(entity.external_id, entity.plugin)
|
|
|
|
] = e
|
|
|
|
if entity.name and entity.plugin:
|
|
|
|
self._entities_cache['by_name_and_plugin'][
|
|
|
|
(entity.name, entity.plugin)
|
|
|
|
] = e
|
|
|
|
|
2022-04-12 00:41:20 +02:00
|
|
|
def _populate_entity_id_from_cache(self, new_entity: Entity):
|
2022-04-11 00:01:21 +02:00
|
|
|
with self._entities_cache_lock:
|
|
|
|
cached_entity = self._get_cached_entity(new_entity)
|
2022-04-12 00:41:20 +02:00
|
|
|
if cached_entity and cached_entity.get('id'):
|
|
|
|
new_entity.id = cached_entity['id']
|
2022-04-11 00:01:21 +02:00
|
|
|
if new_entity.id:
|
|
|
|
self._cache_entities(new_entity)
|
|
|
|
|
|
|
|
def _init_entities_cache(self):
|
|
|
|
with self._get_db().get_session() as session:
|
|
|
|
entities = session.query(Entity).all()
|
|
|
|
for entity in entities:
|
|
|
|
make_transient(entity)
|
|
|
|
|
|
|
|
with self._entities_cache_lock:
|
|
|
|
self._cache_entities(*entities, overwrite_cache=True)
|
|
|
|
|
|
|
|
self.logger.info('Entities cache initialized')
|
|
|
|
|
|
|
|
def _process_event(self, entity: Entity):
|
2022-04-12 00:41:20 +02:00
|
|
|
self._populate_entity_id_from_cache(entity)
|
|
|
|
if entity.id:
|
2022-04-11 00:01:21 +02:00
|
|
|
get_bus().post(EntityUpdateEvent(entity=entity))
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
def post(self, *entities: Entity):
|
|
|
|
for entity in entities:
|
|
|
|
self._queue.put(entity)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def should_stop(self) -> bool:
|
|
|
|
return self._should_stop.is_set()
|
|
|
|
|
|
|
|
def stop(self):
|
|
|
|
self._should_stop.set()
|
|
|
|
|
|
|
|
def run(self):
|
|
|
|
super().run()
|
|
|
|
self.logger.info('Started entities engine')
|
2022-04-11 00:01:21 +02:00
|
|
|
self._init_entities_cache()
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
while not self.should_stop:
|
|
|
|
msgs = []
|
|
|
|
last_poll_time = time()
|
|
|
|
|
|
|
|
while not self.should_stop and (
|
2022-04-05 21:17:58 +02:00
|
|
|
time() - last_poll_time < self._queue_timeout
|
|
|
|
):
|
2022-04-04 16:50:17 +02:00
|
|
|
try:
|
|
|
|
msg = self._queue.get(block=True, timeout=0.5)
|
|
|
|
except Empty:
|
|
|
|
continue
|
|
|
|
|
|
|
|
if msg:
|
|
|
|
msgs.append(msg)
|
2022-04-11 00:01:21 +02:00
|
|
|
# Trigger an EntityUpdateEvent if there has
|
|
|
|
# been a change on the entity state
|
|
|
|
self._process_event(msg)
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
if not msgs or self.should_stop:
|
|
|
|
continue
|
|
|
|
|
2022-04-05 23:04:57 +02:00
|
|
|
try:
|
|
|
|
self._process_entities(*msgs)
|
|
|
|
except Exception as e:
|
|
|
|
self.logger.error('Error while processing entity updates: ' + str(e))
|
|
|
|
self.logger.exception(e)
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
self.logger.info('Stopped entities engine')
|
|
|
|
|
2022-04-05 21:17:58 +02:00
|
|
|
def _get_if_exist(
|
|
|
|
self, session: Session, entities: Iterable[Entity]
|
|
|
|
) -> Iterable[Entity]:
|
2022-04-04 16:50:17 +02:00
|
|
|
existing_entities = {
|
|
|
|
(entity.external_id or entity.name, entity.plugin): entity
|
2022-04-05 21:17:58 +02:00
|
|
|
for entity in session.query(Entity)
|
|
|
|
.filter(
|
|
|
|
or_(
|
|
|
|
*[
|
|
|
|
and_(
|
|
|
|
Entity.external_id == entity.external_id,
|
|
|
|
Entity.plugin == entity.plugin,
|
|
|
|
)
|
|
|
|
if entity.external_id is not None
|
|
|
|
else and_(
|
|
|
|
Entity.name == entity.name, Entity.plugin == entity.plugin
|
|
|
|
)
|
|
|
|
for entity in entities
|
|
|
|
]
|
|
|
|
)
|
|
|
|
)
|
|
|
|
.all()
|
2022-04-04 16:50:17 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
return [
|
|
|
|
existing_entities.get(
|
|
|
|
(entity.external_id or entity.name, entity.plugin), None
|
2022-04-05 21:17:58 +02:00
|
|
|
)
|
|
|
|
for entity in entities
|
2022-04-04 16:50:17 +02:00
|
|
|
]
|
|
|
|
|
|
|
|
def _merge_entities(
|
2022-04-05 21:17:58 +02:00
|
|
|
self, entities: List[Entity], existing_entities: List[Entity]
|
2022-04-04 16:50:17 +02:00
|
|
|
) -> List[Entity]:
|
2022-04-05 21:17:58 +02:00
|
|
|
def merge(entity: Entity, existing_entity: Entity) -> Entity:
|
2022-04-05 23:04:19 +02:00
|
|
|
columns = [col.key for col in entity.columns]
|
2022-04-05 21:17:58 +02:00
|
|
|
for col in columns:
|
|
|
|
if col not in ('id', 'created_at'):
|
|
|
|
setattr(existing_entity, col, getattr(entity, col))
|
|
|
|
|
|
|
|
return existing_entity
|
|
|
|
|
2022-04-04 16:50:17 +02:00
|
|
|
new_entities = []
|
2022-04-05 21:17:58 +02:00
|
|
|
entities_map = {}
|
2022-04-04 16:50:17 +02:00
|
|
|
|
2022-04-05 21:17:58 +02:00
|
|
|
# Get the latest update for each ((id|name), plugin) record
|
|
|
|
for e in entities:
|
|
|
|
key = ((e.external_id or e.name), e.plugin)
|
|
|
|
entities_map[key] = e
|
|
|
|
|
|
|
|
# Retrieve existing records and merge them
|
2022-04-04 16:50:17 +02:00
|
|
|
for i, entity in enumerate(entities):
|
|
|
|
existing_entity = existing_entities[i]
|
|
|
|
if existing_entity:
|
2022-04-05 21:17:58 +02:00
|
|
|
entity = merge(entity, existing_entity)
|
2022-04-04 16:50:17 +02:00
|
|
|
|
2022-04-05 21:17:58 +02:00
|
|
|
new_entities.append(entity)
|
2022-04-04 16:50:17 +02:00
|
|
|
|
|
|
|
return new_entities
|
|
|
|
|
|
|
|
def _process_entities(self, *entities: Entity):
|
2022-04-11 00:01:21 +02:00
|
|
|
with self._get_db().get_session() as session:
|
2022-04-04 16:50:17 +02:00
|
|
|
existing_entities = self._get_if_exist(session, entities)
|
|
|
|
entities = self._merge_entities(entities, existing_entities) # type: ignore
|
|
|
|
session.add_all(entities)
|
|
|
|
session.commit()
|
2022-04-07 01:46:37 +02:00
|
|
|
|
2022-04-11 00:01:21 +02:00
|
|
|
with self._entities_cache_lock:
|
|
|
|
for entity in entities:
|
|
|
|
self._cache_entities(entity, overwrite_cache=True)
|