forked from M-Labs/artiq
118 lines
3.7 KiB
Python
Executable File
118 lines
3.7 KiB
Python
Executable File
#!/usr/bin/env python3.5
|
|
|
|
import asyncio
|
|
import argparse
|
|
import atexit
|
|
import os
|
|
|
|
from artiq.tools import *
|
|
from artiq.protocols.pc_rpc import Server as RPCServer
|
|
from artiq.protocols.sync_struct import Publisher
|
|
from artiq.protocols.logging import Server as LoggingServer
|
|
from artiq.master.log import log_args, init_log
|
|
from artiq.master.databases import DeviceDB, DatasetDB
|
|
from artiq.master.scheduler import Scheduler
|
|
from artiq.master.worker_db import get_last_rid
|
|
from artiq.master.experiments import FilesystemBackend, GitBackend, ExperimentDB
|
|
|
|
|
|
def get_argparser():
|
|
parser = argparse.ArgumentParser(description="ARTIQ master")
|
|
|
|
simple_network_args(parser, [
|
|
("notify", "notifications", 3250),
|
|
("control", "control", 3251),
|
|
("logging", "remote logging", 1066)
|
|
])
|
|
|
|
group = parser.add_argument_group("databases")
|
|
group.add_argument("--device-db", default="device_db.pyon",
|
|
help="device database file (default: '%(default)s')")
|
|
group.add_argument("--dataset-db", default="dataset_db.pyon",
|
|
help="dataset file (default: '%(default)s')")
|
|
|
|
group = parser.add_argument_group("repository")
|
|
group.add_argument(
|
|
"-g", "--git", default=False, action="store_true",
|
|
help="use the Git repository backend")
|
|
group.add_argument(
|
|
"-r", "--repository", default="repository",
|
|
help="path to the repository (default: '%(default)s')")
|
|
|
|
log_args(parser)
|
|
|
|
return parser
|
|
|
|
|
|
def main():
|
|
args = get_argparser().parse_args()
|
|
log_buffer = init_log(args)
|
|
if os.name == "nt":
|
|
loop = asyncio.ProactorEventLoop()
|
|
asyncio.set_event_loop(loop)
|
|
else:
|
|
loop = asyncio.get_event_loop()
|
|
atexit.register(loop.close)
|
|
|
|
device_db = DeviceDB(args.device_db)
|
|
dataset_db = DatasetDB(args.dataset_db)
|
|
dataset_db.start()
|
|
atexit_register_coroutine(dataset_db.stop)
|
|
|
|
if args.git:
|
|
repo_backend = GitBackend(args.repository)
|
|
else:
|
|
repo_backend = FilesystemBackend(args.repository)
|
|
experiment_db = ExperimentDB(repo_backend, device_db.get_device_db)
|
|
atexit.register(experiment_db.close)
|
|
experiment_db.scan_repository_async()
|
|
|
|
worker_handlers = {
|
|
"get_device_db": device_db.get_device_db,
|
|
"get_device": device_db.get,
|
|
"get_dataset": dataset_db.get,
|
|
"update_dataset": dataset_db.update
|
|
}
|
|
scheduler = Scheduler(get_last_rid() + 1, worker_handlers, experiment_db)
|
|
worker_handlers.update({
|
|
"scheduler_submit": scheduler.submit,
|
|
"scheduler_delete": scheduler.delete,
|
|
"scheduler_request_termination": scheduler.request_termination,
|
|
"scheduler_get_status": scheduler.get_status
|
|
})
|
|
scheduler.start()
|
|
atexit_register_coroutine(scheduler.stop)
|
|
|
|
bind = bind_address_from_args(args)
|
|
|
|
server_control = RPCServer({
|
|
"master_device_db": device_db,
|
|
"master_dataset_db": dataset_db,
|
|
"master_schedule": scheduler,
|
|
"master_experiment_db": experiment_db
|
|
})
|
|
loop.run_until_complete(server_control.start(
|
|
bind, args.port_control))
|
|
atexit_register_coroutine(server_control.stop)
|
|
|
|
server_notify = Publisher({
|
|
"schedule": scheduler.notifier,
|
|
"devices": device_db.data,
|
|
"datasets": dataset_db.data,
|
|
"explist": experiment_db.explist,
|
|
"log": log_buffer.data
|
|
})
|
|
loop.run_until_complete(server_notify.start(
|
|
bind, args.port_notify))
|
|
atexit_register_coroutine(server_notify.stop)
|
|
|
|
server_logging = LoggingServer()
|
|
loop.run_until_complete(server_logging.start(
|
|
bind, args.port_logging))
|
|
atexit_register_coroutine(server_logging.stop)
|
|
|
|
loop.run_forever()
|
|
|
|
if __name__ == "__main__":
|
|
main()
|