Run openEO R UDFs from Python
Project description
openeo-udf-python-to-r / openeo-r-udf
This is an experimental engine for openEO to run R UDFs from an R environment.
This currently is limited to R UDFs that are running without any other processes in the following processes:
apply
apply_dimension
reduce_dimension
This repository contains the following content:
- The scripts to run for testing:
tests/test.py
(single core) andtests/test_parallel.py
(parallelized). - The folder
tests/udfs
contains UDF examples as users could provide them. udf_lib.py
is a Python library with the Python code required to run R UDFs from Pythonexecutor.R
is the R script that is run from R and executes the R UDF in the Python environment.
The following image shows how the implementation roughly works:
Install from pypi
This is for back-end developers or end-users who want to test their UDFs locally
You can install this library from pypi:
pip install openeo-r-udf
The following variables should be defined:
udf
(string - The content of the parameterudf
fromrun_udf
, i.e. UDF code or a path/URL to a UDF)udf_folder
(string - The folder where the UDFs reside or should be written to)process
(string - The parent process, i.e.apply
,apply_dimension
orreduce_dimension
)data
(xarray.DataArray - The data to process)dimension
(string, defaults toNone
- The dimension to work on if applicable, doesn't apply forapply
)context
(Any, defaults toNone
- The data that has been passed in thecontext
parameter)
Use it from Python without parallelization:
# import the UDF library
from openeo_r_udf.udf_lib import prepare_udf, execute_udf
# Define variables as documented above
# Load UDF file (this should not be paralelized)
udf_path = prepare_udf(udf, udf_folder)
# Execute UDF file (this can be parallelized)
result = execute_udf(process, udf_path, data, dimension=dimension, context=context)
Use it from Python with parallelization:
# import the UDF library - make sure to install joblib before
from openeo_r_udf.udf_lib import prepare_udf, execute_udf, chunk_cube, combine_cubes
from joblib import Parallel, delayed as joblibDelayed
# Parallelization config
chunk_size = 1000
num_jobs = -1
# Define variables as documented above
# Load UDF file (this should not be paralelized)
udf_path = prepare_udf(udf, udf_folder)
# Define callback function
def compute_udf(data):
return execute_udf(process, udf_path, data.compute(), dimension=dimension, context=context)
# Run UDF in parallel
input_data_chunked = chunk_cube(data, size=chunk_size)
results = Parallel(n_jobs=num_jobs, verbose=51)(joblibDelayed(compute_udf)(data) for data in input_data_chunked)
result = combine_cubes(results)
The result
variable holds the processed data as an xarray.DataArray
again.
Writing a UDF
This is for end-users
A UDF must be written differently depending on where it is executed.
The underlying library used for data handling is always stars
.
apply
A UDF that is executed inside the process apply
manipulates the values on a per-pixel basis.
You can't add or remove labels or dimensions.
The UDF function must be named udf
and receives two parameters:
-
x
is a multi-dimensional stars object and you can run vectorized functions on a per pixel basis, e.g.abs
. -
context
passes through the data that has been passed to thecontext
parameter of the parent process (here:apply
). If nothing has been provided explicitly, the parameter is set toNULL
.This can be used to pass through configurable options, parameters or some additional data. For example, if you would execute
apply(process = run_udf(...), context = list(m = -1, max = -100))
then you could access the corresponding values in the UDF below ascontext$m
andcontext$max
(see example below).
The UDF must return a stars object with exactly the same shape.
Example:
udf = function(x, context) {
max(abs(x) * context$a, context$max)
}
apply_dimension
A UDF that is executed inside the process apply_dimension
takes all values along a dimension and computes new values based on them.
This could for example compute a moving average over a timeseries.
There are two different variants of UDFs that can be used as processes for apply_dimension
.
A reducer can be executed either "vectorized" or "per chunk".
This is the same behavior as defined for reduce_dimension
.
Please see below for more details.
reduce_dimension
A UDF that is executed inside the process reduce_dimension
takes all values along a dimension and computes a single value for it.
This could for example compute an average for a timeseries.
There are two different forms of UDFs that can be used as reducers
for reduce_dimension
: a reducer can be executed either "vectorized"
or "per chunk".
vectorized
The vectorized variant is usually the more efficient variant as it's executed once on a larger chunk of the data cube.
The UDF function must be named udf
and receives two parameters:
data
is a matrix. Each row contains the values for a "pixel" and the columns are the values along the given dimension. So if you reduce along the temporal dimension, the columns are the individual timestamps.context
-> see the description ofcontext
forapply
.
The UDF must return a list of values.
Example:
Please note that you may need to use do.call
to execute functions in a vectorized way. We also need to use pmax
for this, instead of max
.
udf = function(data, context) {
# To get the labels for the values once:
# labels = names(data)
do.call(pmax, data) * context
}
The input data may look like this if you reduce along a band dimension with three bands r
, g
and b
:
data
could belist(r = c(1, 2, 6), g = c(3, 4, 5), b = c(7, 1, 0))
names(data)
would bec("r", "g", "b")
- Executing the example above would return
c(7, 4, 6)
per chunk
This variant is usually slower, but might be required for certain use cases. It is executed multiple times on the smallest chunk possible for the dimension given, e.g. a single time series.
The UDF function must be named udf_chunked
and receives two parameters:
data
is a list of values, e.g. a single time series which you could pass tomax
ormean
.context
-> see the description ofcontext
forapply
.
The UDF must return a single value.
Example:
udf_chunked = function(data, context) {
# To get the labels for the values:
# labels = names(data)
max(data)
}
The input data may look like this if you reduce along a band dimension with three bands r
, g
and b
:
data
could bec(1, 2, 3)
names(data)
would bec("r", "g", "b")
- Exeucting the example above would return
3
Setup and Teardown
As udf_chunked
is usually executed many times in a row, you can optionally define two additional functions that are executed once before and once after the execution.
These functions must be named udf_setup
and udf_teardown
and be placed in the same file as udf_chunked
.
udf_setup
could be useful to initially load some data, e.g. a machine learning (ML) model.
udf_teardown
could be used to clean-up stuff that has been opened in udf_setup
.
Note: udf_setup
and udf_teardown
are only available if you implement udf_chunked
.
If you implement udf
, the two additional functions are not available as you can execute them directly in the udf
function, which is only executed once (for each worker).
Both functions receive a single parameter, which is the context
parameter explained above.
Here the context parameter could contain the path to a ML model file, for example.
By using the context parameter, you can avoid hard-coding information, which helps to make UDFs reusable.
Example:
udf_setup = function(context) {
# e.g. load a ML model from a file
}
udf_chunked = function(data, context) {
max(data)
}
udf_teardown = function(context) {
# e.g. clean-up tasks
}
Note: udf_teardown
is only executed if none of the udf_chunked
calls has resulted in an error.
If you'd like to make some data available in udf_chunked
and/or udf_teardown
that you have prepared in udf_setup
(or udf_chunked
), you can use a global variable
and the special assignment operator <<-
to assign to variables in the outer environments.
Example:
This loads a trained ML model object from an URL in udf_setup
and makes it available for prediction in udf_chunked
.
This is important as loading the ML model in udf_chunked may download the model very often, usually thousands of times and as such the computation gets very slow.
model <- NULL
udf_setup = function(context) {
model <<- load_model("https://example.com/model")
}
udf_chunked = function(data, context) {
return(predict(data, model))
}
Examples
Dockerimage for running on a backend
https://github.com/Open-EO/r4openeo-usecases/tree/main/vito-docker
Implementation at Eurac
tbd
R4openEO use cases
https://github.com/Open-EO/r4openeo-usecases
Development
Clone this repository and switch into the corresponding folder.
- Install environment via conda:
conda env create -f environment.yml
- Install package for development:
pip install -e .
- Now you can run one of the tests for example:
python3 tests/test.py
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file openeo_r_udf-0.4.0.tar.gz
.
File metadata
- Download URL: openeo_r_udf-0.4.0.tar.gz
- Upload date:
- Size: 11.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 413812d85012991bacc151e38434b13c39df698751d7838f62f9226bc542e896 |
|
MD5 | d07a25aa6bd03a39de2cdc3310ea12d9 |
|
BLAKE2b-256 | 8d8873aa68d68bc6833db24e2979e14069f5ff4a96f20d3542ccc864d88d5597 |
File details
Details for the file openeo_r_udf-0.4.0-py3-none-any.whl
.
File metadata
- Download URL: openeo_r_udf-0.4.0-py3-none-any.whl
- Upload date:
- Size: 8.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.7
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | d63d16e4e5713d53ace2d71bdc5b742c87dfce7a1df8c129e0955add0a814f12 |
|
MD5 | e23c53f375bb78fbe1cc46e530fe1444 |
|
BLAKE2b-256 | bd953bf7c8713cbec9950e0e37ebeb98f3a2864f61901fdf25141d818d730890 |