Project.py 5.1 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165
  1. from json import load
  2. from os import path, mkdir
  3. from uuid import uuid1
  4. from eventlet import spawn_after
  5. from pycs.observable import ObservableDict
  6. from pycs.pipeline.PipelineManager import PipelineManager
  7. from pycs.projects.ImageFile import ImageFile
  8. from pycs.projects.VideoFile import VideoFile
  9. from pycs.util.RecursiveDictionary import set_recursive
  10. class Project(ObservableDict):
  11. DEFAULT_PIPELINE_TIMEOUT = 120
  12. def __init__(self, obj: dict, parent):
  13. self.pipeline_manager = None
  14. self.quit_pipeline_thread = None
  15. # ensure all required object keys are available
  16. for key in ['data', 'labels', 'jobs']:
  17. if key not in obj.keys():
  18. obj[key] = {}
  19. # load model data
  20. folder = path.join('projects', obj['id'], 'model')
  21. with open(path.join(folder, 'distribution.json'), 'r') as file:
  22. model = load(file)
  23. model['path'] = folder
  24. obj['model'] = model
  25. # save data as MediaFile objects
  26. for key in obj['data'].keys():
  27. obj['data'][key] = self.create_media_file(obj['data'][key], obj['id'])
  28. # initialize super
  29. super().__init__(obj, parent)
  30. # create data and temp
  31. data_path = path.join('projects', self['id'], 'data')
  32. if not path.exists(data_path):
  33. mkdir(data_path)
  34. temp_path = path.join('projects', self['id'], 'temp')
  35. if not path.exists(temp_path):
  36. mkdir(temp_path)
  37. # subscribe to changes to write to disk afterwards
  38. self.subscribe(lambda d, k: self.parent.write_project(self['id']))
  39. def update_properties(self, update):
  40. set_recursive(update, self)
  41. def new_media_file_path(self):
  42. return path.join('projects', self['id'], 'data'), str(uuid1())
  43. def create_media_file(self, file, project_id=None):
  44. if project_id is None:
  45. project_id = self['id']
  46. # TODO check file extension
  47. # TODO determine type
  48. # TODO filter supported types
  49. if file['extension'] in ['.jpg', '.png']:
  50. return ImageFile(file, project_id)
  51. if file['extension'] in ['.mp4']:
  52. return VideoFile(file, project_id)
  53. raise NotImplementedError
  54. def add_media_file(self, uuid, name, extension, size, created):
  55. file = {
  56. 'id': uuid,
  57. 'name': name,
  58. 'extension': extension,
  59. 'size': size,
  60. 'created': created
  61. }
  62. self['data'][file['id']] = self.create_media_file(file)
  63. def remove_media_file(self, file_id):
  64. del self['data'][file_id]
  65. def add_label(self, name):
  66. label_uuid = str(uuid1())
  67. self['labels'][label_uuid] = {
  68. 'id': label_uuid,
  69. 'name': name
  70. }
  71. def update_label(self, identifier, name):
  72. if identifier in self['labels']:
  73. self['labels'][identifier]['name'] = name
  74. def remove_label(self, identifier):
  75. # abort if identifier is unknown
  76. if identifier not in self['labels']:
  77. return
  78. # remove label from data elements
  79. remove = list()
  80. for data in self['data']:
  81. for pred in self['data'][data]['predictionResults']:
  82. if 'label' in self['data'][data]['predictionResults'][pred]:
  83. if self['data'][data]['predictionResults'][pred]['label'] == identifier:
  84. remove.append((data, pred))
  85. for t in remove:
  86. del self['data'][t[0]]['predictionResults'][t[1]]
  87. # remove label from list
  88. del self['labels'][identifier]
  89. def predict(self, identifiers):
  90. # create pipeline
  91. pipeline = self.__create_pipeline()
  92. # run jobs
  93. for file_id in identifiers:
  94. if file_id in self['data'].keys():
  95. pipeline.run(self['data'][file_id])
  96. # schedule timeout thread
  97. self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
  98. def fit(self):
  99. # create pipeline
  100. pipeline = self.__create_pipeline()
  101. # run fit
  102. pipeline.fit()
  103. # schedule timeout thread
  104. self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
  105. def __create_pipeline(self):
  106. # abort pipeline termination
  107. self.__quit_pipeline_thread()
  108. # create pipeline if it does not exist already
  109. if self.pipeline_manager is None:
  110. self.pipeline_manager = PipelineManager(self)
  111. return self.pipeline_manager
  112. def __quit_pipeline(self):
  113. if self.pipeline_manager is not None:
  114. self.pipeline_manager.close()
  115. self.pipeline_manager = None
  116. self.quit_pipeline_thread = None
  117. def __create_quit_pipeline_thread(self):
  118. # abort pipeline termination
  119. self.__quit_pipeline_thread()
  120. # create new thread
  121. self.quit_pipeline_thread = spawn_after(self.DEFAULT_PIPELINE_TIMEOUT, self.__quit_pipeline)
  122. def __quit_pipeline_thread(self):
  123. if self.quit_pipeline_thread is not None:
  124. self.quit_pipeline_thread.cancel()
  125. self.quit_pipeline_thread = None