Ros Knowledge Network
Project description
See the Homepage for a better formatted version.
Overview
Ros executes graphs of queries to cooperatively compose knowledge networks.
While the language provides common constructs supporting variables, modularity, extensibility, templates, and a type system, it is targeted at the distinctive challenges of creating highly detailed knowledge graphs enabling reasoning and inference.
Usage
Running a workflow locally from the command line produces output like this:
And builds this knowledge graph:
Language Reference
Overview
A workflow is a series of steps. Steps can reference the output of previous steps. In general they have access to a shared graph. They can also exchange sub-graphs. Steps can have associated metadata describing their allowed input and output types.
Workflows compute a directed acyclic graph (DAG) modeling job dependencies which are then executed in the order indicated by a topological sort of the DAG.
Variables
Variables passed to the workflow at the command line or via the API can be resolved dynamically. In this example, $disease_name refers to an argument provided by the execution context to this workflow. The provided value will be substituted at runtime.
In the example below, disease_identifiers
is the job's name. It produces a graph. The JSON object representing that graph will be stored as the value of the job's name. The graph is also written to the shared graph database. Subsequent jobs can interact with either representation.
The code
tag tells the engine which block of functionality to execute.
The args
section lists inputs to this operator.
disease_identifiers:
doc: |
Resolve an English disease name to an ontology identifier.
code: bionames
args:
type: disease
input: $disease_name
Operators
The workflow is organized around graph operators. Each has access to all facilities of the Ros framework including the shared graph.
In general, workflow jobs have the following fields:
- doc: A documentation string.
- code: The name of a component providing functionality.
- args: Arguments to the operator.
Ros currently provides the following core operators:
- get: Invokes an HTTP GET operation on a specified resource.
- union: Unions two or more results into one object.
It also includes certain Translator specific modules. In the future, these will be implemented as Ros plugins:
- biothings: BioThings modules. Currently modules 4 and 5 of workflow 1.
- gamma: Invokes the Gamma reasoner. The example below calls Gamma a few times with different machine questions. It will be updated to use the new Quick API for added flexibility.
- xray: XRay reasoner modules. Currently modules 1 and 2 of workflow 1.
Graphs
Ros provides graphs in two basic modalities:
- Results: Results of previous workflow steps can be referenced as variables passed to subsequent steps. This graph can be queried using JSON Path query syntax. The following example uses the json select facility in the Ros framework to query a variable called condition:
diseases = event.select ( query = "$.[*].result_list.[*].[*].result_graph.node_list.[*]", obj = event.conditions)
- Shared: A shared graph, accessible with the Cypher query language is available to all operators. This example uses the Ros framework's cypher query on the shared graph with biolink-model concepts.
diseases = event.context.graph.query ("match (a:disease) return a")
Each operator receives an event object provided by the Ros framework. The event provides framework services including the shared graph, graph manipulation tools, and arguments to the invocation of the operator.
These facilities allow the operator to query the graphs before executing their main logic and to update it as well.
Metadata
The language supports a metadata capability to enable modules to specify their inputs and outputs.
Inputs support these tags:
- doc: Documentation
- type: Types are currently derived from a standard library but in the future will be extensible and composable.
- required: Whether or not the argument is required.
Outputs support these tags:
- doc: Documentation.
- type: Type of the returned object.
The use of metadata is optional.
Templates
Templates allow extension of the language by specializing existing library functions into new capabilities through composition. Templates are defined in a template section separate from the workflow proper. They can also be defined in separate, reusable modules.
Modules
External modules can be loaded via the import
tag.
A library path like those featured in other high level programming languages governs where libraries are loaded from.
Automated Validation
When a workflow has run, we'd like to run validations that test the integrity of the answer we got. In this first Ros version, we support a limited form of automated validation. The limitations are motivated by making Ros secure. Since we execute queries remotely, it's not realistic to run arbitrary graph queries posted from clients on the internet. Instead, we provide a limited query syntax that still allows us to do a lot of validation. The query syntax is called Syfur - a less capable query system reminiscent of Cypher.
The example below from workflow_one shows Syfur's usage.
automated_validation:
doc: |
Automated Validation
We can test the knowledge network with constraints including all, match, and none.
The depends arg ensures the validation runs only after the graph is built.
Each test has a name, documentation, a query, and an operator defining the constraint to enforce
Syfur
A principal virtue of Ros is the ability to remotely execute workflows.
To make this possible, it must be secured against remote code execution attacks
This includes cypher injection.
It may be possible to accomplish some of this in other ways, but for now, we provide Syfur
Syfur is a restricted DSL that translates to a subset of Cypher. See examples below.
code: validate
args:
depends: $biothings_module_4_and_5
tests:
test_type_2_diabetes_exists:
doc: Ensure type 2 diabetes mellitus and other conditions are present.
items: "match type=disease return id"
all: [ "OMIM:600001", "MONDO:0005148", "DOID:9352" ]
test_pioglitazone:
doc: Test attributes of a specific chemical. Ensure type 2 diabetes mellitus is present.
items: "match type=chemical_substance id=CHEMBL.COMPOUND:CHEMBL595 return node_attributes"
match:
- .*Thiazolidinedione.*
- ".*{'name': 'EPC', 'value': 'Peroxisome Proliferator Receptor gamma Agonist'}.*"
test_absence_of_something:
doc: Verify there is none of something in our graph.
items: "match return id"
none: [ 'BAD:123' ]
Putting it All Together
Here's a usage example to put all of this in context.
1. Define A Template
We begin with the common use case of converting user supplied text into ontology identifiers. We create a template task for this purpose.
- It extends the builtin
get
http operator and sets thepattern
argument to a defined value. - It's saved to a file called
bionames.ros
in a directory that's on the library path.
The bionames template is now a reusable component that can be imported into a variety of workflows.
2. Optionally Model Input and Output Types
Though currently optional, the template also specifies
- The
meta
tag describes metadata about the operator. - The
main
operator is used when no sub-operators are specified. - Each operator has input and output sections.
- The input section specifies a list of input values.
- Each may have
doc
,type
, andrequired
tags. - The output section may contain
doc
andtype
tags. - In both cases, values of
type
must (currently) come from the Ros standard library, described elsewhere.
doc: |
This module defines a reusable template.
It can be imported into other workflows via the import directive.
templates:
bionames:
doc: |
This template extends the built in get operator.
code: get
args:
pattern: 'https://bionames.renci.org/lookup/{input}/{type}/'
meta:
main:
args:
input:
doc: The name of the object to find identifiers for.
type: string
required: true
type:
doc: The intended biolink_model type of the resulting identifiers.
type: biolink_model
required: true
output:
doc: A Translator standard knowledge graph.
type: knowledge_graph_standard
The biolink_model
and knowledge_graph_standard
types are currently modeled directly in the Ros standard library:
types:
string :
doc: A primitive string of characters.
extends: primitive
biolink_model:
doc: An element from the Biolink-model
extends: string
knowledge_graph_standard:
doc: A Translator knowledge graph standard (KGS) knowledge graph.
extends: primitive
3. Build the Workflow
Next, we import the template above into a workflow definition.
doc: |
NCATS Biomedical Translator - Workflow One
import:
- bionames
workflow:
disease_identifiers:
doc: |
Resolve an English disease name to an ontology identifier.
code: bionames
args:
type: disease
input: $disease_name
...
The disease_identifiers
task instantiates the bionames template. It supplies the required arguments, including resolving the disease_name
input argument specific to this workflow.
Once this module has executed, subsequent steps can reference the graph it produced via the variable $disease_identifiers
.
The next step in the workflow makes use of this by referencing the output of our template invocation. The workflow engine sees this reference and makes the new job depend on the referenced job. This is how the job DAG governing execution order is calculated.
condition_to_drug:
doc: |
Module 1
* What are the defining symptoms / phenotypes of [condition x]?
* What conditions present [symptoms]?
* Filter [conditions] to only keep ones with defined genetic causes.
* What subset of conditions are most representative of [conditions]? (find archetypes)
Module 2
* What genes are implicated in [condition]?
* What subset of genes are most representative of [conditions]? (find archetypes)
* What pathways/processes are [genes] involved in?
* What genes are involved in [pathway/process]?
* What drugs/compounds target gene products of [gene]?
Invoke XRay module 1 and 2 given the disease identifier from bionames.
The graph argument references the entire bionames response.
The op argument specifies the XRay operation to execute.
code: xray
args:
op: condition_expansion_to_gene_pathway_drug
graph: $disease_identifiers
For more details, see the whole workflow.
Execution
For all workflows, the engine builds a directed acyclic graph (DAG) of jobs by examining each job to determine its dependencies.
A topological sort of the jobs provides the execution order.
Execution is asynchronous. To the extent you subscribe to the semantic distinction between 'concurrent' and 'parallel' workflow execution is concurrent but not parallel.
The current implementation uses Python's asyncio. This means multiple operations are able to make progress during the same time window. But it does not mean that the operations literally execute (run CPU operations) simultaneously. The current profile of operations is I/O bound rather than CPU bound so this approach is likely to be enough for a while. Several tasks can wait for an HTTP request to return while others use the processor to handle results.
The API also handles requests asynchronously using Sanic and Tornado
Getting Started
Docker
Requirements
- Docker
- Docker Compose (included with Docker on Mac)
- Git
- Ports 7474, 7687, and 5002 available
Start
git clone git@github.com:NCATS-Tangerine/ros.git
cd ros/deploy
source ros.env
docker-compose up
Usage - Command Line
- Connect to the local Neo4J
- Connect to the API docker container
- Run a workflow via the API.
$ cd ../ros
$ PYTHONPATH=$PWD/.. python app.py --api --workflow workflows/workflow_one.ros -l workflows -i disease_name="type 2 diabetes mellitus" --out stdout
Usage - Programmatic
Ros can execute workflows remotely and return the resulting knowledge network. The client currently supports JSON and NetowrkX representations.
from ros.client import Client
ros = Client (url="http://localhost:5002")
response = ros.run (workflow='workflows/workflow_one.ros',
args = { "disease_name" : "type 2 diabetes mellitus" },
library_path = [ 'workflows' ])
graph = response.to_nx ()
for n in graph.nodes (data=True):
print (n)
Install
Requirements:
- Python >3.7.x
- Neo4J >3.3.4
Steps:
These steps install the package, print help text, and execute workflow one. To run this, you'll need workflow_one.ros and bionames.ros from the repo. The -l workflows
flag names the directory containing the workflows.
$ pip install ros
$ ros --help
$ ros --workflow workflow_one.ros --arg disease_name="diabetes mellitus type 2" -l workflows
NDEx
To save a workflow to NDEx.
-
Create an NDEx account.
-
Create an ~/.ndex credential file like this:
{ "username" : "<username>", "password" : "<password>" }
-
Run the workflow with NDEx output:
ros flow --workflow workflow_one.ros --out output.json --ndex wf1
Help
$ PYTHONPATH=$PWD/.. python app.py --help
usage: app.py [-h] [-a] [-w WORKFLOW] [-s SERVER] [-i ARG] [-o OUT]
[-l LIBPATH] [-n NDEX] [--validate]
Ros Workflow CLI
optional arguments:
-h, --help show this help message and exit
-a, --api URL of the remote Ros server to use.
(default: False)
-w WORKFLOW, --workflow WORKFLOW Workflow to execute. (default:
workflow_one.ros)
-s SERVER, --server SERVER Hostname of api server (default:
http://localhost:5002)
-i ARG, --arg ARG Add an argument expressed as key=val
(default: [])
-o OUT, --out OUT Output the workflow result graph to a
file. Use 'stdout' to print to terminal.
(default: None)
-l LIBPATH, --libpath LIBPATH A directory containing workflow modules.
(default: ['.'])
-n NDEX, --ndex NDEX Name of the graph to publish to NDEx.
Requires valid ~/.ndex credential file.
(default: None)
--validate Validate inputs and outputs (default:
False)
Next
- Information Architecture: Develop:
- Controlled vocabulary: Especially regarding what the modules are and how they relate
- Input and Output Signatures: For the modules
- Provenance: Both in terms of workflow provenance (which user, how long, etc) and metadata about sources (SCEPIO?).
- Polymorphism: It would be really helpful if multiple entities implementing a capability could implement the same OpenAPI interface to enable polymorphic invocation. This would also help with parallelism.
- KGX: Maybe KGX should be the shared graph, at least optionally. Just need to design that connection.
- Concurrent / Parallel / Distributed: Ros now supports concurrent task execution via Python async. If this turns out not to be enough, explore running via something capable of parallel and maybe distributed execution.
- Composability: Allow workflows to import and reuse other workflows.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.