forked from M-Labs/artiq
master: optimize repository scan, closes #546
This commit is contained in:
parent
4ef5eb2644
commit
387688354c
|
@ -2,6 +2,7 @@ import asyncio
|
||||||
import os
|
import os
|
||||||
import tempfile
|
import tempfile
|
||||||
import shutil
|
import shutil
|
||||||
|
import time
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
from artiq.protocols.sync_struct import Notifier
|
from artiq.protocols.sync_struct import Notifier
|
||||||
|
@ -13,16 +14,19 @@ from artiq.tools import get_windows_drives, exc_to_warning
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
async def _get_repository_entries(entry_dict,
|
class _RepoScanner:
|
||||||
root, filename, worker_handlers):
|
def __init__(self, worker_handlers):
|
||||||
worker = Worker(worker_handlers)
|
self.worker_handlers = worker_handlers
|
||||||
|
self.worker = None
|
||||||
|
|
||||||
|
async def process_file(self, entry_dict, root, filename):
|
||||||
|
logger.debug("processing file %s %s", root, filename)
|
||||||
try:
|
try:
|
||||||
description = await worker.examine("scan", os.path.join(root, filename))
|
description = await self.worker.examine(
|
||||||
|
"scan", os.path.join(root, filename))
|
||||||
except:
|
except:
|
||||||
log_worker_exception()
|
log_worker_exception()
|
||||||
raise
|
raise
|
||||||
finally:
|
|
||||||
await worker.close()
|
|
||||||
for class_name, class_desc in description.items():
|
for class_name, class_desc in description.items():
|
||||||
name = class_desc["name"]
|
name = class_desc["name"]
|
||||||
arginfo = class_desc["arginfo"]
|
arginfo = class_desc["arginfo"]
|
||||||
|
@ -46,8 +50,7 @@ async def _get_repository_entries(entry_dict,
|
||||||
}
|
}
|
||||||
entry_dict[name] = entry
|
entry_dict[name] = entry
|
||||||
|
|
||||||
|
async def _scan(self, root, subdir=""):
|
||||||
async def _scan_experiments(root, worker_handlers, subdir=""):
|
|
||||||
entry_dict = dict()
|
entry_dict = dict()
|
||||||
for de in os.scandir(os.path.join(root, subdir)):
|
for de in os.scandir(os.path.join(root, subdir)):
|
||||||
if de.name.startswith("."):
|
if de.name.startswith("."):
|
||||||
|
@ -55,19 +58,28 @@ async def _scan_experiments(root, worker_handlers, subdir=""):
|
||||||
if de.is_file() and de.name.endswith(".py"):
|
if de.is_file() and de.name.endswith(".py"):
|
||||||
filename = os.path.join(subdir, de.name)
|
filename = os.path.join(subdir, de.name)
|
||||||
try:
|
try:
|
||||||
await _get_repository_entries(
|
await self.process_file(entry_dict, root, filename)
|
||||||
entry_dict, root, filename, worker_handlers)
|
|
||||||
except Exception as exc:
|
except Exception as exc:
|
||||||
logger.warning("Skipping file '%s'", filename,
|
logger.warning("Skipping file '%s'", filename,
|
||||||
exc_info=not isinstance(exc, WorkerInternalException))
|
exc_info=not isinstance(exc, WorkerInternalException))
|
||||||
|
# restart worker
|
||||||
|
await self.worker.close()
|
||||||
|
self.worker = Worker(self.worker_handlers)
|
||||||
if de.is_dir():
|
if de.is_dir():
|
||||||
subentries = await _scan_experiments(
|
subentries = await self._scan(
|
||||||
root, worker_handlers,
|
root, os.path.join(subdir, de.name))
|
||||||
os.path.join(subdir, de.name))
|
|
||||||
entries = {de.name + "/" + k: v for k, v in subentries.items()}
|
entries = {de.name + "/" + k: v for k, v in subentries.items()}
|
||||||
entry_dict.update(entries)
|
entry_dict.update(entries)
|
||||||
return entry_dict
|
return entry_dict
|
||||||
|
|
||||||
|
async def scan(self, root):
|
||||||
|
self.worker = Worker(self.worker_handlers)
|
||||||
|
try:
|
||||||
|
r = await self._scan(root)
|
||||||
|
finally:
|
||||||
|
await self.worker.close()
|
||||||
|
return r
|
||||||
|
|
||||||
|
|
||||||
def _sync_explist(target, source):
|
def _sync_explist(target, source):
|
||||||
for k in list(target.read.keys()):
|
for k in list(target.read.keys()):
|
||||||
|
@ -109,7 +121,9 @@ class ExperimentDB:
|
||||||
self.repo_backend.release_rev(self.cur_rev)
|
self.repo_backend.release_rev(self.cur_rev)
|
||||||
self.cur_rev = new_cur_rev
|
self.cur_rev = new_cur_rev
|
||||||
self.status["cur_rev"] = new_cur_rev
|
self.status["cur_rev"] = new_cur_rev
|
||||||
new_explist = await _scan_experiments(wd, self.worker_handlers)
|
t1 = time.monotonic()
|
||||||
|
new_explist = await _RepoScanner(self.worker_handlers).scan(wd)
|
||||||
|
logger.info("repository scan took %d seconds", time.monotonic()-t1)
|
||||||
|
|
||||||
_sync_explist(self.explist, new_explist)
|
_sync_explist(self.explist, new_explist)
|
||||||
finally:
|
finally:
|
||||||
|
|
|
@ -76,6 +76,8 @@ class Worker:
|
||||||
return "worker({},{})".format(self.rid, self.filename)
|
return "worker({},{})".format(self.rid, self.filename)
|
||||||
|
|
||||||
async def _create_process(self, log_level):
|
async def _create_process(self, log_level):
|
||||||
|
if self.ipc is not None:
|
||||||
|
return # process already exists, recycle
|
||||||
await self.io_lock.acquire()
|
await self.io_lock.acquire()
|
||||||
try:
|
try:
|
||||||
if self.closed.is_set():
|
if self.closed.is_set():
|
||||||
|
|
Loading…
Reference in New Issue