fix: unused options in redis fix: compatibility between python 2-3 fix: forgot vars declarations fix: carry conditions in listing redis keys fix: listing redis list keys fix: removed duplicated tasks when they are listed fix: index number when redis DB are listed fix: some error levels in log fix: renamed *proc* -> *tasks* files fix: added the process manager backend for 'tasks' options, thinking in future to add new process managers
84 lines
1.9 KiB
Python
84 lines
1.9 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
import six
|
|
import csv
|
|
import logging
|
|
|
|
from time import sleep
|
|
from kombu import Connection
|
|
|
|
from .utils import list_remote_process
|
|
|
|
log = logging.getLogger()
|
|
|
|
|
|
# ----------------------------------------------------------------------
|
|
def action_proc_raw_dump(config):
|
|
|
|
log.warning(" - Trying to connect with server...")
|
|
|
|
url = '%s://%s' % (config.broker_type, config.target)
|
|
|
|
f_output = None
|
|
csv_output = None
|
|
|
|
if config.output is not None:
|
|
fixed_f = "%s.csv" % config.output if ".csv" not in config.output else config.output
|
|
|
|
f_output = open(fixed_f, "a")
|
|
csv_output = csv.writer(f_output)
|
|
|
|
log.error(" - Storing results at '%s'" % fixed_f)
|
|
|
|
# Write first col
|
|
csv_output.writerow([
|
|
"# Task name",
|
|
"Parameters (position#value)"
|
|
])
|
|
|
|
already_processed = set()
|
|
|
|
# with Connection('redis://%s' % REDIS) as conn:
|
|
with Connection(url) as conn:
|
|
in_queue = conn.SimpleQueue('celery')
|
|
|
|
while 1:
|
|
|
|
for remote_task, remote_args, task_id in list_remote_process(config, in_queue):
|
|
|
|
# Task already processed?
|
|
if task_id not in already_processed:
|
|
|
|
# Track
|
|
already_processed.add(task_id)
|
|
|
|
# Show info
|
|
log.error(" Found process information:")
|
|
log.error(" - Remote tasks name: '%s'" % remote_task)
|
|
log.error(" - Input parameters:")
|
|
|
|
to_csv = [remote_task]
|
|
|
|
for i, x in enumerate(remote_args):
|
|
log.error(" -> P%s: %s" % (i, x))
|
|
|
|
# Prepare to store JSON
|
|
to_csv.append("%s#%s" % (i, x))
|
|
|
|
# Store
|
|
if csv_output is not None:
|
|
csv_output.writerow(to_csv)
|
|
|
|
# Queue is empty -> wait
|
|
if config.streaming_mode:
|
|
log.error(" -> No more messages from server. Waiting for %s seconds and try again.." % config.interval)
|
|
sleep(config.interval)
|
|
else:
|
|
log.error(" -> No more messages from server. Exiting...")
|
|
return
|
|
|
|
# Close file descriptor
|
|
if f_output is not None:
|
|
f_output.close()
|
|
csv_output.close()
|