forked from mercure-imaging/mercure
-
Notifications
You must be signed in to change notification settings - Fork 0
/
router.py
executable file
·215 lines (177 loc) · 7.71 KB
/
router.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
"""
router.py
=========
mercure's central router module that evaluates the routing rules and decides which series should be sent to which target.
"""
# Standard python includes
import asyncio
import time
import signal
import os
import sys
import uuid
import graphyte
import daiquiri
import hupper
from typing import Dict
# App-specific includes
from common.constants import mercure_defs, mercure_names
import common.helper as helper
import common.config as config
import common.monitor as monitor
from routing.route_series import route_series, route_error_files
from routing.route_studies import route_studies
from routing.common import generate_task_id
import common.influxdb
import common.notification as notification
# Create local logger instance
logger = config.get_logger()
main_loop = None # type: helper.AsyncTimer # type: ignore
async def terminate_process(signalNumber, frame) -> None:
"""
Triggers the shutdown of the service
"""
helper.g_log("events.shutdown", 1)
logger.info("Shutdown requested")
monitor.send_event(monitor.m_events.SHUTDOWN_REQUEST, monitor.severity.INFO)
# Note: main_loop can be read here because it has been declared as global variable
if "main_loop" in globals() and main_loop.is_running:
main_loop.stop()
helper.trigger_terminate()
def run_router() -> None:
"""
Main processing function that is called every second
"""
if helper.is_terminated():
return
helper.g_log("events.run", 1)
# logger.info('')
# logger.info('Processing incoming folder...')
try:
config.read_config()
except Exception:
logger.warning( # handle_error
"Unable to update configuration. Skipping processing.",
None,
event_type=monitor.m_events.CONFIG_UPDATE,
)
return
filecount = 0
series: Dict[str, float] = {}
complete_series: Dict[str, float] = {}
pending_series: Dict[str, float] = {} # Every series that hasn't timed out yet
error_files_found = False
# Check the incoming folder for completed series. To this end, generate a map of all
# series in the folder with the timestamp of the latest DICOM file as value
for entry in os.scandir(config.mercure.incoming_folder):
if entry.name.endswith(mercure_names.TAGS) and not entry.is_dir():
filecount += 1
seriesString = entry.name.split(mercure_defs.SEPARATOR, 1)[0]
modificationTime = entry.stat().st_mtime
if seriesString in series.keys():
if modificationTime > series[seriesString]:
series[seriesString] = modificationTime
else:
series[seriesString] = modificationTime
# Check if at least one .error file exists. In that case, the incoming folder should
# be searched for .error files at the end of the update run
if (not error_files_found) and entry.name.endswith(mercure_names.ERROR):
error_files_found = True
# Check if any of the series exceeds the "series complete" threshold
for series_entry in series:
if (time.time() - series[series_entry]) > config.mercure.series_complete_trigger:
complete_series[series_entry] = series[series_entry]
else:
pending_series[series_entry] = series[series_entry]
# logger.info(f'Files found = {filecount}')
# logger.info(f'Series found = {len(series)}')
# logger.info(f'Complete series = {len(complete_series)}')
helper.g_log("incoming.files", filecount)
helper.g_log("incoming.series", len(series))
# Process all complete series
for series_uid in sorted(complete_series):
task_id = generate_task_id()
try:
route_series(task_id, series_uid)
except Exception:
logger.error(f"Problems while processing series {series_uid}", task_id) # handle_error
# If termination is requested, stop processing series after the active one has been completed
if helper.is_terminated():
return
if error_files_found:
route_error_files()
# Now, check if studies in the studies folder are ready for routing/processing
route_studies(pending_series)
def exit_router(args) -> None:
"""
Callback function that is triggered when the process terminates. Stops the asyncio event loop
"""
helper.loop.call_soon_threadsafe(helper.loop.stop)
# Main entry point of the router module
def main(args=sys.argv[1:]) -> None:
if "--reload" in args or os.getenv("MERCURE_ENV", "PROD").lower() == "dev":
# start_reloader will only return in a monitored subprocess
reloader = hupper.start_reloader("router.main")
logger.info("")
logger.info(f"mercure DICOM Router ver {mercure_defs.VERSION}")
logger.info("--------------------------------------------")
logger.info("")
# Register system signals to be caught
signals = (signal.SIGTERM, signal.SIGINT)
for s in signals:
helper.loop.add_signal_handler(s, lambda s=s: asyncio.create_task(terminate_process(s, helper.loop)))
instance_name = "main"
# Read the optional instance name from the argument (if running multiple instances in one appliance)
if len(sys.argv) > 1:
instance_name = sys.argv[1]
# Read the configuration file and terminate if it cannot be read
try:
config.read_config()
except Exception:
logger.exception("Cannot start service. Going down.")
sys.exit(1)
appliance_name = config.mercure.appliance_name
logger.info(f"Appliance name = {appliance_name}")
logger.info(f"Instance name = {instance_name}")
logger.info(f"Instance PID = {os.getpid()}")
logger.info(sys.version)
notification.setup()
monitor.configure("router", instance_name, config.mercure.bookkeeper)
monitor.send_event(monitor.m_events.BOOT, monitor.severity.INFO, f"PID = {os.getpid()}")
if len(config.mercure.graphite_ip) > 0:
logger.info(f"Sending events to graphite server: {config.mercure.graphite_ip}")
graphite_prefix = "mercure." + appliance_name + ".router." + instance_name
graphyte.init(config.mercure.graphite_ip, config.mercure.graphite_port, prefix=graphite_prefix)
if len(config.mercure.influxdb_host) > 0:
logger.info(f"Sending events to influxdb server: {config.mercure.influxdb_host}")
common.influxdb.init(
config.mercure.influxdb_host,
config.mercure.influxdb_token,
config.mercure.influxdb_org,
config.mercure.influxdb_bucket,
"mercure." + appliance_name + ".router." + instance_name
)
logger.info(
f"""Incoming folder: {config.mercure.incoming_folder}
Studies folder: {config.mercure.studies_folder}
Outgoing folder: {config.mercure.outgoing_folder}
Processing folder: {config.mercure.processing_folder}"""
)
# Start the timer that will periodically trigger the scan of the incoming folder
global main_loop
main_loop = helper.AsyncTimer(config.mercure.router_scan_interval, run_router)
helper.g_log("events.boot", 1)
try:
main_loop.run_until_complete(helper.loop)
# Process will exit here once the asyncio loop has been stopped
monitor.send_event(monitor.m_events.SHUTDOWN, monitor.severity.INFO)
except Exception as e:
monitor.send_event(monitor.m_events.SHUTDOWN, monitor.severity.ERROR, str(e))
finally:
# Finish all asyncio tasks that might be still pending
remaining_tasks = helper.asyncio.all_tasks(helper.loop) # type: ignore[attr-defined]
if remaining_tasks:
helper.loop.run_until_complete(helper.asyncio.gather(*remaining_tasks))
logger.info("Going down now")
if __name__ == "__main__":
main()