This is intended for developers only, see README for the main docs.
For running locally without Docker you will need
- robot
- owltools
- python3.6 or higher
See Dockerfile for details on how to obtain these
Previously ODK used a perl script to create a new repo. This iterated the template/ directory and used special magic for expanding into a target folder. This has been replaced by python code odk/odk.py with makes used of Jinja2 templates.
For example, the file
template/src/ontology/Makefile.jinja2
will compile to a file src/ontology/Makefile
in the target/output
directory.
Jinja2 templates should be fairly easy to grok for anyone familiar with templating systems. The syntax is very similar to Liquid templates, which are used extensively on the OBO site. We feed the template engine with a project object that is passed in by the user (more on that later).
Logic in the templates should be non-existent.
Sometimes the odk needs to create a file whose name is based on an input setting or configuration; sometimes lists of such files need to be created.
For example, if the user specifies 3 external ontology dependencies,
then we want to see the repo with 3 files imports/{{ont.id}}_import.owl
Rather than embed this logic in code, we include all dynamic files in a single "tar-esque" formated file: template/_dynamic_files.jinja2
This file is actually a specification for multiple files, each target
file specified with ^^^
. Because the parent file is interpreted
using templates, we can have dynamic file names, and entire files
created via looping constructs.
Currently the datamodel is specified as python dataclasses, for now
the best way to see the complete spec is to look at the classes
annotated with @dataclass
in the code.
There is a schema folder but this is incomplete as the dataclasses-scheme module doesn't appear to work (TODO)...
There are also example project.yaml
files in the
examples folder, and these also serve as rudimentary unit
tests.
See for example examples/go-mini/project.yaml
The basic data model is:
- An
OntologyProject
consists of various configuration settings, plusProductGroup
s - These are:
- An
ImportProduct
group which specifies how import files are generated - A
SubsetProduct
group which specifies how subset/slim files are generated - Other product groups for reports and templates
- An
Many ontology projects need only specify a very minimal configuration: id of ontology, github/gitlab location, and list of ontology ids for imports. However, for projects that need to customize there are multiple options. E.g. for an import product you can optionally specific a particular URL that overrides the default PURL.
Note that for backwards compatibility, a project.yaml file is not
required. A user can specify an entire repo by running seed
with
options such as -d
for dependencies.
Note that in all cases a project.yaml
file is generated.
$ ./odk/odk.py --help
Usage: odk.py [OPTIONS] COMMAND [ARGS]...
Options:
--help Show this message and exit.
Commands:
create-dynfile For testing purposes
create-makefile For testing purposes
dump-schema Dumps the python schema as json schema.
export-project For testing purposes
seed Seeds an ontology project
The most common command is seed.
Previously with odk there was no path to either upgrading an existing project with new settings (i.e. adding an import) OR to take advantage of changes to the odk (e.g changes in the core Makefile).
This should now be easier with the new odk, although the implementation emphasis has been on the seed command. Some things that will make this easier:
- Convention of using a second loaded Makefile for custom changes
- Maintaining a project.yaml in root folder will allow easy regeneration
TODO: add a refresh command. This could run odk in place, but preserving protected files. TBD how to determine protected files. Obviously the edit file should not be touched. Could use git log to determine if any modifications have been made?
Note that with v1.2 the main odkfull Dockerfile is at the root level. We now use a base alpine image for compactness, and selectively add in unix tools like make and rsync.
Note also that we include odk.py and the template folders in the image. This means that odk seed can now be run from anywhere!
To build the Docker image from the top level:
make docker-build
Note that this means local invocations to use obolibrary/odkfull
will use the version you built.
To test:
make docker-test
To publish on Dockerhub:
make docker-publish
There is a potential for some confusion as to responsibility for logic. On the one hand we have dependency logic in the Makefile. But we also have minimal logic in deciding what to put in the Makefile.
For example, we could move some logic from the Makefile by using for/endfor Jinja constructs and unfolding every product in a group and have an explicit non-pattern target in the Makefile. Or we can continue to write targets with patterns. Or we can do a mixture of both.
Additionally there is some minimal logic in the python odk code, but this is kept to an absolute minimum; the role of the python code is to run template expansions.
In general the decision is to keep the templating as simple as possible, which leads to a slight mixed two level system.
One gotcha is the two levels of comments. The {# .. #}
comments are
template comments for the eyes of developers only. These are ignored
when compiling down to the target file. Then we also have Makefile
comments #
which remain in the target file, and are intended for
advanced ontology maintainers who need to debug their workflows. These
are intermingled in Makefile.jinja2
To run:
make test
These will seed a few example repos in the target/ folder, some from command line opts, others from a project.yaml
These are pseudo-tests as the output is not examined, however they do serve to guard against multiple kinds of errors as the seed script will often fail if things are not set up correctly.
The examples folder serves for both unit test and documentation purposes.
TODO