forked from M-Labs/artiq
master: optimize repository scan, closes #546
This commit is contained in:
parent
f23ebc035b
commit
ad697d4320
|
@ -2,6 +2,7 @@ import asyncio
|
|||
import os
|
||||
import tempfile
|
||||
import shutil
|
||||
import time
|
||||
import logging
|
||||
|
||||
from artiq.protocols.sync_struct import Notifier
|
||||
|
@ -13,16 +14,19 @@ from artiq.tools import get_windows_drives, exc_to_warning
|
|||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
async def _get_repository_entries(entry_dict,
|
||||
root, filename, worker_handlers):
|
||||
worker = Worker(worker_handlers)
|
||||
class _RepoScanner:
|
||||
def __init__(self, worker_handlers):
|
||||
self.worker_handlers = worker_handlers
|
||||
self.worker = None
|
||||
|
||||
async def process_file(self, entry_dict, root, filename):
|
||||
logger.debug("processing file %s %s", root, filename)
|
||||
try:
|
||||
description = await worker.examine("scan", os.path.join(root, filename))
|
||||
description = await self.worker.examine(
|
||||
"scan", os.path.join(root, filename))
|
||||
except:
|
||||
log_worker_exception()
|
||||
raise
|
||||
finally:
|
||||
await worker.close()
|
||||
for class_name, class_desc in description.items():
|
||||
name = class_desc["name"]
|
||||
arginfo = class_desc["arginfo"]
|
||||
|
@ -46,8 +50,7 @@ async def _get_repository_entries(entry_dict,
|
|||
}
|
||||
entry_dict[name] = entry
|
||||
|
||||
|
||||
async def _scan_experiments(root, worker_handlers, subdir=""):
|
||||
async def _scan(self, root, subdir=""):
|
||||
entry_dict = dict()
|
||||
for de in os.scandir(os.path.join(root, subdir)):
|
||||
if de.name.startswith("."):
|
||||
|
@ -55,19 +58,28 @@ async def _scan_experiments(root, worker_handlers, subdir=""):
|
|||
if de.is_file() and de.name.endswith(".py"):
|
||||
filename = os.path.join(subdir, de.name)
|
||||
try:
|
||||
await _get_repository_entries(
|
||||
entry_dict, root, filename, worker_handlers)
|
||||
await self.process_file(entry_dict, root, filename)
|
||||
except Exception as exc:
|
||||
logger.warning("Skipping file '%s'", filename,
|
||||
exc_info=not isinstance(exc, WorkerInternalException))
|
||||
# restart worker
|
||||
await self.worker.close()
|
||||
self.worker = Worker(self.worker_handlers)
|
||||
if de.is_dir():
|
||||
subentries = await _scan_experiments(
|
||||
root, worker_handlers,
|
||||
os.path.join(subdir, de.name))
|
||||
subentries = await self._scan(
|
||||
root, os.path.join(subdir, de.name))
|
||||
entries = {de.name + "/" + k: v for k, v in subentries.items()}
|
||||
entry_dict.update(entries)
|
||||
return entry_dict
|
||||
|
||||
async def scan(self, root):
|
||||
self.worker = Worker(self.worker_handlers)
|
||||
try:
|
||||
r = await self._scan(root)
|
||||
finally:
|
||||
await self.worker.close()
|
||||
return r
|
||||
|
||||
|
||||
def _sync_explist(target, source):
|
||||
for k in list(target.read.keys()):
|
||||
|
@ -109,7 +121,9 @@ class ExperimentDB:
|
|||
self.repo_backend.release_rev(self.cur_rev)
|
||||
self.cur_rev = new_cur_rev
|
||||
self.status["cur_rev"] = new_cur_rev
|
||||
new_explist = await _scan_experiments(wd, self.worker_handlers)
|
||||
t1 = time.monotonic()
|
||||
new_explist = await _RepoScanner(self.worker_handlers).scan(wd)
|
||||
logger.info("repository scan took %d seconds", time.monotonic()-t1)
|
||||
|
||||
_sync_explist(self.explist, new_explist)
|
||||
finally:
|
||||
|
|
|
@ -76,6 +76,8 @@ class Worker:
|
|||
return "worker({},{})".format(self.rid, self.filename)
|
||||
|
||||
async def _create_process(self, log_level):
|
||||
if self.ipc is not None:
|
||||
return # process already exists, recycle
|
||||
await self.io_lock.acquire()
|
||||
try:
|
||||
if self.closed.is_set():
|
||||
|
|
Loading…
Reference in New Issue