123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135 |
- from json import load
- from os import path
- from uuid import uuid1
- from eventlet import spawn_after
- from pycs.observable import ObservableDict
- from pycs.pipeline.PipelineManager import PipelineManager
- from pycs.projects.MediaFile import MediaFile
- from pycs.util.RecursiveDictionary import set_recursive
- class Project(ObservableDict):
- DEFAULT_PIPELINE_TIMEOUT = 120
- def __init__(self, obj: dict, parent):
- self.pipeline_manager = None
- self.quit_pipeline_thread = None
- # ensure all required object keys are available
- for key in ['data', 'labels', 'jobs']:
- if key not in obj.keys():
- obj[key] = {}
- # load model data
- folder = path.join('projects', obj['id'], 'model')
- with open(path.join(folder, 'distribution.json'), 'r') as file:
- model = load(file)
- model['path'] = folder
- obj['model'] = model
- # save data as MediaFile objects
- for key in obj['data'].keys():
- obj['data'][key] = MediaFile(obj['data'][key], obj['id'])
- # initialize super
- super().__init__(obj, parent)
- # subscribe to changes to write to disk afterwards
- self.subscribe(lambda d, k: self.parent.write_project(self['id']))
- def update_properties(self, update):
- set_recursive(update, self)
- def new_media_file_path(self):
- return path.join('projects', self['id'], 'data'), str(uuid1())
- def add_media_file(self, file):
- file = MediaFile(file, self['id'])
- self['data'][file['id']] = file
- def remove_media_file(self, file_id):
- del self['data'][file_id]
- def add_label(self, name):
- label_uuid = str(uuid1())
- self['labels'][label_uuid] = {
- 'id': label_uuid,
- 'name': name
- }
- def update_label(self, identifier, name):
- if identifier in self['labels']:
- self['labels'][identifier]['name'] = name
- def remove_label(self, identifier):
- # abort if identifier is unknown
- if identifier not in self['labels']:
- return
- # remove label from data elements
- remove = list()
- for data in self['data']:
- for pred in self['data'][data]['predictionResults']:
- if 'label' in self['data'][data]['predictionResults'][pred]:
- if self['data'][data]['predictionResults'][pred]['label'] == identifier:
- remove.append((data, pred))
- for t in remove:
- del self['data'][t[0]]['predictionResults'][t[1]]
- # remove label from list
- del self['labels'][identifier]
- def predict(self, identifiers):
- # create pipeline
- pipeline = self.__create_pipeline()
- # run jobs
- for file_id in identifiers:
- if file_id in self['data'].keys():
- pipeline.run(self['data'][file_id])
- # schedule timeout thread
- self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
- def fit(self):
- # create pipeline
- pipeline = self.__create_pipeline()
- # run fit
- pipeline.fit()
- # schedule timeout thread
- self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
- def __create_pipeline(self):
- # abort pipeline termination
- self.__quit_pipeline_thread()
- # create pipeline if it does not exist already
- if self.pipeline_manager is None:
- self.pipeline_manager = PipelineManager(self)
- return self.pipeline_manager
- def __quit_pipeline(self):
- if self.pipeline_manager is not None:
- self.pipeline_manager.close()
- self.pipeline_manager = None
- self.quit_pipeline_thread = None
- def __create_quit_pipeline_thread(self):
- # abort pipeline termination
- self.__quit_pipeline_thread()
- # create new thread
- self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
- def __quit_pipeline_thread(self):
- if self.quit_pipeline_thread is not None:
- self.quit_pipeline_thread.cancel()
- self.quit_pipeline_thread = None
|