forked from rucio/rucio
-
Notifications
You must be signed in to change notification settings - Fork 0
/
rucio-conveyor-submitter
executable file
·94 lines (86 loc) · 4.14 KB
/
rucio-conveyor-submitter
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
#!/usr/bin/env python
# Copyright 2013-2018 CERN for the benefit of the ATLAS collaboration.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# Authors:
# - Mario Lassnig, <[email protected]>, 2013-2015
# - Cedric Serfon, <[email protected]>, 2013
# - Wen Guan, <[email protected]>, 2014-2016
# - Vincent Garonne, <[email protected]>, 2016-2018
"""
Conveyor is a daemon to manage file transfers.
"""
import argparse
import signal
from rucio.daemons.conveyor.submitter import run, stop
def get_parser():
"""
Returns the argparse parser.
"""
parser = argparse.ArgumentParser()
parser.add_argument("--run-once", action="store_true", default=False,
help='One iteration only')
parser.add_argument("--process", action="store", default=0, type=int,
help='Concurrency control: current processes number')
parser.add_argument("--total-processes", action="store", default=1, type=int,
help='Concurrency control: total number of processes')
parser.add_argument("--total-threads", action="store", default=1, type=int,
help='Concurrency control: total number of threads per process')
parser.add_argument("--bulk", action="store", default=100, type=int,
help='Bulk control: number of requests')
parser.add_argument("--group-bulk", action="store", default=1, type=int,
help='Group control: number of requests per group')
parser.add_argument("--group-policy", action="store", default='rule', type=str,
help='Group control: policy used to group. enum{rule, dest, src_dest, rule_src_dest}')
parser.add_argument("--mock", action="store_true", default=False,
help='Fake source replicas for requests')
parser.add_argument('--fts-source-strategy', action="store", default='auto', type=str,
help='FTS source strategy')
parser.add_argument('--exclude-rses', action="store", default=None, type=str,
help='RSE expression to exclude')
parser.add_argument('--include-rses', action="store", default=None, type=str,
help='RSE expression to include')
parser.add_argument('--rses', nargs='+', type=str,
help='Explicit list of RSEs to include')
parser.add_argument('--activities', nargs='+', type=str,
help='Explicit list of activities to include')
parser.add_argument('--sleep-time', action="store", default=600, type=int,
help='Seconds to sleep if few requests')
parser.add_argument('--max-sources', action="store", default=4, type=int,
help='Maximum source replicas per FTS job')
parser.add_argument("--retry-other-fts", action="store_true", default=False,
help='retry on a different FTS')
return parser
if __name__ == "__main__":
signal.signal(signal.SIGTERM, stop)
parser = get_parser()
args = parser.parse_args()
try:
run(once=args.run_once,
process=args.process,
total_processes=args.total_processes,
total_threads=args.total_threads,
bulk=args.bulk,
group_bulk=args.group_bulk,
mock=args.mock,
include_rses=args.include_rses,
exclude_rses=args.exclude_rses,
rses=args.rses,
fts_source_strategy=args.fts_source_strategy,
activities=args.activities,
sleep_time=args.sleep_time,
max_sources=args.max_sources,
retry_other_fts=args.retry_other_fts)
except KeyboardInterrupt:
stop()