Coordinator - better-wfp-00001 pipe¶
This coordinator processes the better-wfp-00001 data pipeline queue for the Sentinel-1 backscatter timeseries
- First do the imports of the Python libraries required
In [1]:
import sys
import os
import owslib
from owslib.wps import monitorExecution
from owslib.wps import WebProcessingService
import lxml.etree as etree
import json
import cioppy
from shapely.wkt import loads
import getpass
import folium
from datetime import datetime, timedelta
import dateutil.parser
from nbconvert.preprocessors import ExecutePreprocessor, CellExecutionError
import nbformat as nbf
import requests
- Read the data pipeline configuration information:
In [2]:
%store -r
nb_config = os.path.join('..', 'configuration.ipynb')
nb = nbf.read(nb_config, 4)
exec(nb['cells'][1]['source']) in globals(), locals()
app = dict([('artifact_id', app_artifact_id),
('version', app_version),
('repository', repository),
('community', community)])
app_process_id = '%s_%s_%s_%s' % (app['community'].replace('-', '_'),
app['artifact_id'].replace('-', '_'),
app['artifact_id'].replace('-', '_'),
app['version'].replace('.', '_'))
trigger_pipe = dict([('artifact_id', trigger_pipe_artifact_id),
('version', trigger_pipe_version),
('repository', repository),
('folder', folder),
('community', community)])
trigger_pipe_process_id = '%s_%s_%s_%s' % (trigger_pipe['community'].replace('-', '_'),
trigger_pipe['artifact_id'].replace('-', '_'),
trigger_pipe['artifact_id'].replace('-', '_'),
trigger_pipe['version'].replace('.', '_'))
print 'This notebook will process the queue of %s with the trigger %s' % (app_process_id,
trigger_pipe_process_id)
This notebook will process the queue of ec_better_ewf_wfp_01_01_01_ewf_wfp_01_01_01_1_16 with the trigger ec_better_tg_wfp_01_01_01_pipe_tg_wfp_01_01_01_pipe_0_11
Queue selection parameters¶
In [3]:
series = 'https://catalog.terradue.com/%s/series/source-queue/description' % data_pipeline
In [4]:
series
Out[4]:
'https://catalog.terradue.com/better-wfp-00001/series/source-queue/description'
Coordinator parameters¶
In [7]:
coordinator_name = 'co_%s_pipe' % data_pipeline
coordinator_date_start = '2019-01-27T00:00Z'
coordinator_date_stop = '2019-01-27T23:59Z'
coordinator_period = '0 16 * * *'
In [8]:
start_pipe = '${coord:formatTime(coord:dateOffset(coord:nominalTime(), -2, \'HOUR\'), "yyyy-MM-dd\'T\'HH:mm:ss\'Z\'")}'
end_pipe = '${coord:formatTime(coord:dateOffset(coord:nominalTime(), -0, \'HOUR\'), "yyyy-MM-dd\'T\'HH:mm:ss\'Z\'")}'
In [9]:
co_trigger_pipe_process_id = 'coordinator_%s' % trigger_pipe_process_id
Common Parameters¶
In [10]:
tg_quotation = 'No'
recovery = 'No'
_T2Username = data_pipeline
Check data transformation application¶
In [5]:
wps_url_apps = '%s/zoo-bin/zoo_loader.cgi' % apps_deployer
wps = WebProcessingService(wps_url_apps, verbose=False, skip_caps=False)
found_process = False
message = "The process %s is not deployed" % app_process_id
for index, elem in enumerate(wps.processes):
if elem.identifier == app_process_id:
message = "The process %s is deployed" % app_process_id
found_process = True
print message
if not found_process:
raise Exception()
The process ec_better_ewf_wfp_01_01_01_ewf_wfp_01_01_01_1_16 is deployed
Check trigger coordinator¶
In [12]:
wps_url_triggers = '%s/zoo-bin/zoo_loader.cgi' % trigger_deployer
wps = WebProcessingService(wps_url_triggers, verbose=False, skip_caps=False)
found_process = False
message = "The pipe coordinator process %s is not deployed" % co_trigger_pipe_process_id
for index, elem in enumerate(wps.processes):
if elem.identifier == co_trigger_pipe_process_id:
message = "The pipe coordinator process %s is deployed" % co_trigger_pipe_process_id
found_process = True
print message
if not found_process:
raise Exception(message)
The pipe coordinator process coordinator_ec_better_tg_wfp_01_01_01_pipe_tg_wfp_01_01_01_pipe_0_9 is deployed
Process the queue¶
In [13]:
process = wps.describeprocess(co_trigger_pipe_process_id)
print process.title
print process.abstract
WFP-01-01-01 Trigger - Pipe Coordinator
Coordinator: Trigger for the WFP-01-01-01 Sentinel-1 backscatter timeseries data pipeline - Pipe
In [14]:
for data_input in process.dataInputs:
print data_input.identifier
series
data_pipeline
wps_url
process_id
api_key
recovery
t2_coordinator_date_start
t2_coordinator_date_stop
t2_coordinator_period
t2_coordinator_name
quotation
_T2Username
In [15]:
inputs = [('series', series),
('data_pipeline', data_pipeline),
('wps_url', wps_url_apps),
('process_id', app_process_id),
('api_key', datapipeline_api_key),
('recovery', 'No'),
('t2_coordinator_date_start', coordinator_date_start),
('t2_coordinator_date_stop', coordinator_date_stop),
('t2_coordinator_period', coordinator_period),
('t2_coordinator_name', coordinator_name),
('quotation', tg_quotation),
('_T2Username', data_pipeline)]
Submit the coordinator request¶
In [16]:
execution = owslib.wps.WPSExecution(url=wps_url_triggers)
execution_request = execution.buildRequest(co_trigger_pipe_process_id,
inputs,
output=[('coordinatorIds', False)])
execution_response = execution.submitRequest(etree.tostring(execution_request, pretty_print=True))
execution.parseResponse(execution_response)
execution.statusLocation
monitorExecution(execution)
if not execution.isSucceded():
raise Exception('Coordinator %s creation failed' % co_trigger_queue_process_id)
In [17]:
coordinator_id = str(json.loads(execution.processOutputs[0].data[0])['coordinatorsId'][0]['oozieId'])
In [18]:
coordinator_id
Out[18]:
'0028293-180330140554685-oozie-oozi-C'
** DANGER ZONE **
In [16]:
answer = raw_input('Are you sure you want to kill the coordinator %s (YES I DO to confirm)?' % coordinator_id)
if answer == 'YES I DO':
r = requests.put('%s:11000/oozie/v1/job/%s?user.name=oozie&action=kill' % (production_centre, coordinator_id))
if r.status_code:
print 'Coordinator %s killed' % coordinator_id