# manta-ray-client **Repository Path**: mirrors_ICRAR/manta-ray-client ## Basic Information - **Project Name**: manta-ray-client - **Description**: Python API and helper scripts to interact with the MWA ASVO. - **Primary Language**: Unknown - **License**: Not specified - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-08-08 - **Last Updated**: 2026-02-14 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Manta-ray Client (MWA ASVO Command Line Client) ## Description Python API and helper script (mwa_client) to interact with the [MWA ASVO](https://asvo.mwatelescope.org). Giant Squid is the preferred CLI client for the MWA ASVO- check it out here [Giant Squid](https://github.com/MWATelescope/giant-squid) For general help on using the MWA ASVO, please visit: [MWA ASVO wiki](https://mwatelescope.atlassian.net/wiki/spaces/MP/pages/24973129/Data+Access). --- NOTE FOR HPC USERS Please read [this wiki article](https://mwatelescope.atlassian.net/wiki/spaces/MP/pages/65405030/MWA+ASVO+Use+with+HPC+Systems) if you are running manta-ray-client on HPC systems. --- - Supported Python versions: - Python 3.10 - Python 3.9 - Python 3.8 ## mwa_client mwa_client is a helper script which provides the following functions: - Submit MWA ASVO jobs in bulk - Monitor the status of your jobs - Download your completed jobs There are three types of MWA ASVO jobs: - Conversion: Average, convert and download a visibility data set (and optionally apply calibration solutions). - Download: Package and download a raw visibility data set. (This is recommended for advanced users, as the raw visibility files are in an MWA-specific format and require conversion and calibration). - Voltage: Raw voltage data from VCS observations. This option is restricted to members of the mwavcs team who have a Pawsey account. If you are interested in getting access to VCS data, please [contact us](maito:asvo_support@mwatelescope.org) ## Installation Options You must have an account on the [MWA ASVO website](https://asvo.mwatelescope.org) Set your API key as an environment variables in linux (usually in your profile / .bashrc). You can get your API key from [Your Profile page](https://asvo.mwatelescope.org/settings) on the MWA ASVO website. ```bash ~$ export MWA_ASVO_API_KEY= ``` Then you may install natively on your computer OR install via Docker. ### Installation Natively on your computer #### Clone the repository ```bash ~$ git clone https://github.com/ICRAR/manta-ray-client.git ``` #### Create a virtual environment ```bash python3 -m venv env ``` or if you are _still_ using python2.7 you will need to use virtualenv (See [Setting up Python, Pip, and Virtualenv (external link)](http://timsherratt.org/digital-heritage-handbook/docs/python-pip-virtualenv/) for information on installing virtualenv) ```bash ~$ virtualenv -p /usr/bin/python2.7 env ``` #### Activate the virtual environment ```bash ~$ source env/bin/activate (env)~$ ``` #### Install mwa_client and all required packages ```bash (env)~$ cd manta-ray-client (env)~/manta-ray-client$ pip3 install -r requirements.txt (env)~/manta-ray-client$ python3 setup.py install ``` ### Installation using Docker If you prefer, you can also run the manta-ray-client as a Docker container instead of installing it locally. This assumes you have docker installed on your machine. If not please see the [Get Docker (external link)](https://docs.docker.com/get-docker/) page for instructions. #### Clone the repository ```bash ~$ git clone https://github.com/mwatelescope/manta-ray-client.git ``` #### Build the image ```bash ~$ cd manta-ray-client ~/manta-ray-client$ docker build --tag manta-ray-client:latest . ``` #### Use The Container Once the image is built, you can run the mwa_client directly. The below command will: - Create and launch and instances of the image (called a container), - Map '/your/host/data/path/' which should be a directory on your machine, to the container's /data directory - Remove the container once it has finished the command - Map your machine's MWA_ASVO_API_KEY environment variable into the container so it has your MWA ASVO API key - Then 'mwa_client -w all -d /data' will run the mwa_client and download all 'Completed' jobs to the container's /data directory (which we mapped to '/your/host/data/path/' on your machine) ```bash ~$ docker run --name my_mwa_client --entrypoint="" --volume=/your/host/data/path/:/data --rm=true -e MWA_ASVO_API_KEY manta-ray-client:latest mwa_client -w all -d /data ``` Or you can open a shell within the container itself and then run as many mwa_client commands as you like, interactively, then exit to leave the container: ```bash ~$ docker run -it --name my_mwa_client --entrypoint="" --volume=/your/host/data/path/:/data --rm=true -e MWA_ASVO_API_KEY manta-ray-client:latest /bin/bash root@c197566f86d9:/# mwa_client -l ... root@c197566f86d9:/# exit ~$ ``` You will get a prompt like the one above and from there you can run mwa_client commands as normal. ## Examples ```bash mwa_client -c csvfile -d destdir Submit jobs in the csv file, monitor them, then download the files, then exit mwa_client -c csvfile -s Submit jobs in the csv file, then exit mwa_client -d destdir -w JOBID Download the job id (assuming it is ready to download), then exit mwa_client -d destdir -w all Download any ready to download jobs, then exit mwa_client -d destdir -w all -e error_file Download any ready to download jobs, then exit, writing any errors to error_file mwa_client -l List all of your jobs and their status, then exit ``` ## Help ```bash optional arguments: -h, --help Show this help message and exit -s, --submit-only Submit job(s) from csv file then exit (-d is ignored) -l, --list-only List the user's active job(s) and exit immediately (-s, -c & -d are ignored) -w DOWNLOAD_JOB_ID, --download-only DOWNLOAD_JOB_ID Download the job id (-w DOWNLOAD_JOB_ID), if it is ready; or all downloadable jobs (-w all | -w 0), then exit (-s, -c & -l are ignored) -c FILE, --csv FILE csv job file -d DIR, --dir DIR Download directory -e ERRFILE, --error-file ERRFILE, --errfile ERRFILE Write errors in json format to an error file -v, --verbose Verbose output -ar, --allow-resubmit Will allow a job with the same parameters and an existing job in your queue in Completed, Error or Cancelled status to be resubmitted. Default is to not allow resubmission if the new job matches the parameters of an existing job in your queue. ``` ## Job States Each job submitted will transition through the following states: - Queued: Job has been submitted and is waiting to be processed. - Waiting for calibration: New Calibration solution has been requested. - Staging: Files are being staged. - Staged: Files are staged from the Archive. - Retrieving Files from Archive: Files are being downloaded to HPC from the archive. - Preprocessing: Job is being processed by Birli. - Imaging: image product is being created. - Delivering: Files are being delivered to the destined location. - Ready for download: Job has completed- job product is ready for download. - Downloading: Job product is being downloaded. - Download Compete: Product download has been completed. - Error: There was an error. ## Submitting Jobs Users can submit multiple jobs using a CSV file (see below for instructions). ## CSV Format Each row is a single job and each CSV element must be a key=value pair. Whitespace (blank rows) and comments (lines beginning with #) are allowed. Please see the included [example.csv](example.csv) for several full working examples. ## Conversion Job Options Please note that some options are only available depending on the choice of preprocessor (explained below). - `obs_id: ` - Observation ID - `job_type: c` - Always 'c' for conversion jobs. - `avg_time_res: ` - Time resolution: average N seconds of time steps together before writing output. - `avg_freq_res: ` - Average N kHz bandwidth of fine channels together before writing output. - `flag_edge_width: ` - Defaults to 80 kHz. - Flag the given width (in kHz) of edge channels of each coarse channel. - Set to 0 kHz to disable edge flagging. - `output: ` - Output format. - `ms`: CASA measurement set. - `uvfits`: uvfits output. - `delivery: ` - Where you would like your data to be stored - `acacia (default)`: Data will be delivered to Pawsey's Acacia system and you will receive a link to download a zip file containing the data. - `scratch`: Data will be left on the /scratch file system at Pawsey in /scratch/\/asvo/. - `dug`: Data will be transferred to the dug super computing fecility under /data//asvo/. - astro and scratch options are only available for Pawsey users who are in one of the mwa science groups (mwasci, mwavcs, mwaeor, mwaops). Please contact support if you would like to use this option. ### Flags / Optional Options - In addition to the options specified above, a number of flags (or optional options) can also be passed with the job request. - To enable an option, set value to true e.g. `no_rfi=true` - If you omit an option it is equivalent to false. e.g. not specifying no_rfi is equivalent to `no_rfi=false`. Birli currently supports the options below. For more info on the Birli preprocessor, please visit [the repository](https://github.com/mwatelescope/birli). Any other flags passed will be ignored. - `no_rfi=true` Do not perform RFI detection. - `no_geometric_delay=true` Disable geometric corrections. - `no_cable_delay=true` Disable cable length corrections. - `no_digital_gains=true` Do not correct for the digital gains. - `no_passband_gains=true` Apply unity passband (i.e. do not apply any passband corrections). - `no_flag_dc=true` Do not flag the centre/DC channel of each coarse channel. #### Calibration - `apply_di_cal=true` Apply a rough calibration solution to the dataset, if found. If not found, the job will fail- in this case you can resubmit the job without this option for uncalibrated raw visibilities. See: [Data Access/MWA ASVO Calibration Option](https://mwatelescope.atlassian.net/wiki/spaces/MP/pages/24972245/MWA+ASVO+Calibration+Option) on the [MWA Telescope Wiki](https://mwatelescope.atlassian.net/wiki/spaces/MP/overview?mode=global) for more information. #### Pointing options If the `centre` options is omitted, the job will default to using the observations phase centre. - `centre=phase || pointing || custom` - `phase` Centre on the observations phase centre - `pointing` Centre on the observations pointing centre - `custom` Centre on a custom phase centre. If this option is specified, two additional parameters must be passed: - `phase_centre_ra: ` ICRS (J2000.0). Centre on a custom phase centre with this decimal right ascension (must include phase_centre_dec). - `phase_centre_dec: ` ICRS (J2000.0). Centre on a custom phase centre with this decimal declination (must include phase_centre_ra). - e.g. `centre=custom,phase_centre_ra=123.23,phase_centre_dec=-20.1` #### Example line in csv file ```csv obs_id=1110103576, job_type=c, avg_time_res=8, avg_freq_res=40, flag_edge_width=80, output=ms, apply_di_cal=true, no_rfi=true ``` ### Download Job Options - `obs_id: ` - Observation ID - `job_type: d` - Always 'd' for download jobs. - `download_type: ` - `vis_meta`: download visibility metadata only (metafits and RFI flags). - `vis`: download raw visibility data sets and metadata (raw visibility files, metafits and RFI flags). - `delivery: ` - `acacia`: Data will be delivered to Pawsey's Acacia system and you will receive a link to download a zip file containing the data. - `scratch`: Data will be left on the /scratch file system at Pawsey in /scratch/\/asvo/. - `dug`: Data will be transferred to the dug super computing fecility under /data//asvo/. - astro and scratch options are only available for Pawsey users who are in one of the mwa science groups (mwasci, mwavcs, mwaeor, mwaops). Please contact support if you would like to use this option. #### Example lines in csv file ```csv obs_id=1110103576, job_type=d, download_type=vis, delivery=acacia obs_id=1110105120, job_type=d, download_type=vis_meta, delivery=scratch obs_id=1110105120, job_type=d, download_type=vis_meta, delivery=dug ``` ### Voltage Job Options Note that voltage jobs will always be left on /astro or /scratch, and you will therefore need a Pawsey account to submit them. Please get in contact if you're interested in accessing VCS data. - `obs_id: ` - Observation ID - `job_type: v` - Always 'v' for voltage jobs. - `offset: ` - Number of seconds from the beginning of the observation for which you would like data - `duration: ` - Number of seconds of voltage data to be included in the job. #### Example lines in csv file ```csv obs_id=1323776840, job_type=v, offset=0, duration=1200 ``` ### Understanding and using the error file output You can get a machine readable error file in JSON format by specifying "-e" | "--error-file" | "--errfile" on the command line. This might be useful if you are trying to automate the download and processing of many observations and you don't want to try and parse the human readable standard output. An example of the format is below, with two jobs with errors: ```json [ { "obs_id": "1216295963", "job_id": 28979, "result": "Error: an error message" }, { "obs_id": "1216298341", "job_id": 28980, "result": "Error: some error message" } ] ``` Since this is JSON, in python you could simply use the below code to iterate through any errors by deserialising the JSON string: ```python import json # Open the error file mwa_client produced when using -e with open("error.txt", "r") as f: # Read the JSON from the file into a string json_string = f.read() # Deserialise the JSON into a python list of objects result_list = json.loads(json_string) # Iterate through all of the errors for r in result_list: print("Job:{0} ObsId:{1} Result:{2}", r['job_id'], r['obs_id'], r['result']) ```