BioSAXS tasks#
How to use
These tasks come from ewoksbm29. It can be installed with
pip install ewoksbm29
ℹ️ These tasks are used at the following ESRF beamlines: BM29
DahuHplcSummary#
Rebuild the complete chromatogram with basic analysis.
In addition to the inputs from DahuJobWithIspybUpload
:
Required inputs:
integrated_files (List[str]): Azimuthally integrated SAXS data.
Optional inputs:
hplc_summary_parameters (dict): Extra subtract parameters (see
HplcSummaryParameters
).
- Identifier:
ewoksbm29.tasks.hplc_summary.DahuHplcSummary
- Task type:
class
- Required inputs:
integrated_files
- Optional inputs:
config_directory, dahu_job_index, dahu_parameter_file, dahu_url, extra_dahu_parameters, hplc_summary_parameters, ispyb_metadata, ispyb_url, nobackup, timeout
- Outputs:
dahu_job_id, dahu_job_index, dahu_result
DahuIntegrate#
Azimuthal integration of BM29 SAXS data.
In addition to the inputs from DahuJobWithIspybUpload
:
Required inputs:
scan_data_slice (dict): Subset of parameters related to the scan data (see
IntegrateParameters
).
Optional inputs:
integrate_parameters (dict): Extra integrate parameters (see
IntegrateParameters
).
- Identifier:
ewoksbm29.tasks.integrate.DahuIntegrate
- Task type:
class
- Required inputs:
scan_data_slice
- Optional inputs:
config_directory, dahu_job_index, dahu_parameter_file, dahu_url, extra_dahu_parameters, integrate_parameters, ispyb_metadata, ispyb_url, nobackup, timeout
- Outputs:
dahu_job_id, dahu_job_index, dahu_result
ReadScanDataSlice#
Return a subset of IntegrateParameters
for a SAXS scan index range covered
by the lima file with index lima_file_index
.
Optional inputs (online & offline):
lima_file_index (int): Lima file index that determines the scan slice.
dahu_to_counter_name (dict): Map counter names to
IntegrateParameters
keys.storage_ring_current (float): Storage ring current in mA.
exposure_time (float): Scan point exposure time in s.
Optional inputs (online):
scan_key (str): Blissdata scan key (tries offline when fails).
retry_timeout (float): Timeout when trying to access the Lima image dataset or wait for the scan to be PREPARED.
retry_period (float): Period in retry loops.
Optional inputs (offline):
scan_file_path (str): Bliss scan file name.
scan_number (str): Bliss scan number.
Outputs:
scan_data_slice (dict): Subset of
IntegrateParameters
related to the scan data.has_data (bool):
scan_data_slice
contains a non-empty scan data slice.lima_file_index (int): The lime file index to which
scan_data_slice
belongs.next_lima_file_index (int): The next Lima file index.
- Identifier:
ewoksbm29.tasks.read.ReadScanDataSlice
- Task type:
class
- Required inputs:
- Optional inputs:
dahu_to_counter_name, exposure_time, lima_file_index, retry_period, retry_timeout, scan_file_path, scan_key, scan_number, storage_ring_current
- Outputs:
has_data, lima_file_index, next_lima_file_index, scan_data_slice
AccumulateDahuJobResults#
Accumulate Dahu job results with the job index as a key.
.. code: python
dahu_results[dahu_job_index] = {"dahu_job_id": dahu_job_id, "dahu_result": dahu_result}
Required inputs:
dahu_job_index (int): Dahu job index for ordering results.
dahu_job_id (Optional[int]): Dahu job id (
None
means no Dahu job was executed).dahu_result (Optional[dict]): Dahu result (
None
means no Dahu job was executed).
Outputs:
dahu_results (Dict[int, dict]): Add the Dahu result to results from previous executions within the same workflow.
- Identifier:
ewoksbm29.tasks.results.AccumulateDahuJobResults
- Task type:
class
- Required inputs:
dahu_job_id, dahu_job_index, dahu_result
- Optional inputs:
- Outputs:
dahu_results
DahuSubtract#
Subtract average integrated buffer signal from the sample signal.
In addition to the inputs from DahuJobWithIspybUpload
:
Required inputs:
sample_file (str): Azimuthally integrated SAXS data.
buffer_files (List[str]): Azimuthally integrated SAXS data of the buffer.
Optional inputs:
subtract_parameters (dict): Extra subtract parameters (see
SubtractParameters
).
- Identifier:
ewoksbm29.tasks.subtract.DahuSubtract
- Task type:
class
- Required inputs:
buffer_files, sample_file
- Optional inputs:
config_directory, dahu_job_index, dahu_parameter_file, dahu_url, extra_dahu_parameters, ispyb_metadata, ispyb_url, nobackup, subtract_parameters, timeout
- Outputs:
dahu_job_id, dahu_job_index, dahu_result