Jina Hub is an open-registry for hosting immutable Jina components via container images. It enables users to ship, and exchange their best-practice across various Jina search applications.
Features
β Complex dependency? Solved! A clean way to make your deep learning models and fancy algorithms reproducible and accessible everywhere.
π Simply Share - You only need to upload a Python file, a Dockerfile and a YAML. Leave the rest of the work to us. The building, version control is automatically done on the cloud side.
βοΈ Cloud-native - You can easily scale out the images as elastic microservices using Flow API, Docker Swarm, Kubernetes, or your favorite orchestration framework.
π¦ Community-driven - Hub images are made and owned by the community. You can search and follow the image you like, comment, and discuss the problem you met.
Build my own Pod image |
Use Pod image in my project |
Publish Pod image to Jina Hub |
View all images on Jina Hub |
Instead of
jina pod --yaml-path hub/examples/mwu_encoder/mwu_encoder.yml --port-in 55555 --port-out 55556
After this tutorial, you can use the Pod image via:
docker run jinaai/hub.examples.mwu_encoder --port-in 55555 --port-out 55556
...or use the Pod image in the Flow API:
from jina.flow import Flow
f = (Flow()
.add(name='my-encoder', image='jinaai/hub.examples.mwu_encoder', port_in=55555, port_out=55556)
.add(name='my-indexer', yaml_path='indexer.yml'))
... or use the Pod image via Jina CLI
jina pod --image jinaai/hub.examples.mwu_encoder --port-in 55555 --port-out 55556
More information about the usage can be found here.
So you have implemented an awesome executor and want to reuse it in another Jina application, or share it with people in the world. Kind as you are, you want to offer people a ready-to-use interface without hassling them to repeat all steps and pitfalls you have done. The best way is thus to pack everything (python file, YAML config, pre-trained data, dependencies) into a container image and use Jina as the entry point. You can also annotate your image with some meta information to facilitate the search, archive, and classification.
Here are a list of reasons that may motivate you to build a Pod image:
- You want to use one of the built-in executors (e.g. pytorch-based) and you don't want to install pytorch dependencies on the host.
- You modify or write a new executor and want to reuse it in another project, without touching
jina-ai/jina
. - You customize the driver and want to reuse it in another project, without touching
jina-ai/jina
. - You have a self-built library optimized for your architecture (e.g. tensorflow/numpy on GPU/CPU/x64/arm64), and you want this specific Pod to benefit from it.
- Your awesome executor requires certain Linux headers that can only be installed via
apt
oryum
, but you don't havesudo
on the host. - Your executor relies on a pre-trained model, you want to include this 100MB file into the image so that people don't need to download it again.
- You use Kubernetes or Docker Swarm and this orchestration framework requires each microservice to run as a Docker container.
- You are using Jina on the cloud and you want to deploy an immutable Pod and version control it.
- You have figured out a set of parameters that works best for an executor, you want to write it down in a YAML config and share it with others.
- You are debugging, doing try-and-error, or exploring new packages, and you don't want to ruin your local dev environments.
Typically, the following files are packed into the container image:
File | Descriptions |
---|---|
Dockerfile |
describes the dependency setup and expose the entry point; |
manifest.yml |
metadata of the image, author, tags, etc. help the Hub to index and classify your image |
*.py |
describes the executor logic written in Python, if applicable; |
*.yml |
a YAML file describes the executor arguments and configs, if you want users to use your config; |
Other data files | may be required to run the executor, e.g. pre-trained model, fine-tuned model, home-made data. |
Except Dockerfile
, all others are optional to build a valid Pod image depending on your case. manifest.yml
is only required when you want to upload your image to Jina Hub.
In this example, we consider the scenario where we create a new executor and want to reuse it in another project, without touching jina-ai/jina
. All files required in this guide are available at hub/examples/mwu_encoder
.
We write a new dummy encoder named MWUEncoder
in mwu_encoder.py
which encodes any input into a random 3-dimensional vector. This encoder has a dummy parameter greetings
which prints a greeting message on start and on every encode. In mwu_encoder.yml
, the metas.py_modules
tells Jina to load the class of this executor from mwu_encoder.py
.
!MWUEncoder
with:
greetings: im from internal yaml!
metas:
name: my-mwu-encoder
py_modules: mwu_encoder.py
workspace: ./
The documentations of the YAML syntax can be found here.
The Dockerfile
in this example is as simple as three lines,
FROM jinaai/jina:devel
ADD *.py mwu_encoder.yml ./
ENTRYPOINT ["jina", "pod", "--yaml-path", "mwu_encoder.yml"]
Let's now look at these three lines one by one.
FROM jinaai/jina:devel
In the first line, we choose jinaai/jina:devel
as the base image, which corresponds to the latest master of jina-ai/jina
. But of course, you are free to use others, e.g. tensorflow/tensorflow:nightly-gpu-jupyter
.
In practice, whether to use Jina base image depends on the dependencies you would like to introduce. For example, someone provides a hard-to-compile package as a Docker image, much harder than compiling/installing Jina itself. In this case, you may want to use this image as the base image to save some troubles. But don't forget to install Python >=3.7 (if not included) and Jina afterwards, e.g.
FROM awesome-gpu-optimized-kit
RUN pip install jina --no-cache-dir --compile
The ways of installing Jina can be found here.
βΉοΈ In CI pipeline, one can also install Jina from the source via:
ADD jina .src
RUN pip install .src/.
, where jina
is the master branch of jina-ai/jina mounted to the building context by our CI pipeline.
In this example, our dummy MWUEncoder
only requires Jina and does not need any third-party framework. Thus, jinaai/jina:devel
is used.
ADD *.py mwu_encoder.yml ./
The second step is to add all necessary files to the image. Typically, Python codes, YAML config and some data files.
In this example, our dummy MWUEncoder
does not require extra data files.
ENTRYPOINT ["jina", "pod", "--yaml-path", "mwu_encoder.yml"]
The last step is to specify the entrypoint of this image, usually via jina pod
.
In this example, we set mwu_encoder.yml
as a default YAML config. So if the user later run
docker run jinaai/hub.examples.mwu_encoder
It equals to:
jina pod --yaml-path hub/example/mwu_encoder.yml
Any followed key-value arguments after docker run jinaai/hub.examples.mwu_encoder
will be passed to jina pod
. For example,
docker run jinaai/hub.examples.mwu_encoder --port-in 55555 --port-out 55556
It equals to:
jina pod --yaml-path hub/example/mwu_encoder.yml --port-in 55555 --port-out 55556
One can also override the internal YAML config by giving an out-of-docker external YAML config via:
docker run $(pwd)/hub/example/mwu_encoder_ext.yml:/ext.yml jinaai/hub.examples.mwu_encoder --yaml-path /ext.yml
Now you can build the Pod image via docker build
:
cd hub/example
docker build -t jinaai/hub.examples.mwu_encoder .
Depending on whether you want to use the latest Jina image, you may first pull it via docker pull jinaai/jina:devel
before the build. For the sake of immutability, docker build
will not automatically pull the latest image for you.
Congratulations! You can now re-use this Pod image how ever you want.
The most powerful way to use this Pod image is via Docker CLI directly:
docker run --rm -p 55555:55555 -p 55556:55556 jinaai/hub.examples.mwu_encoder --port-in 55555 --port-out 55556
Note, the exposure of ports -p 55555:55555 -p 55556:55556
is required for other Pods (local/remote) to communicate this Pod. One may also want to use --network host
and let the Pod share the network layer of the host.
All parameters supported by jina pod --help
can be followed after docker run jinaai/hub.examples.mwu_encoder
.
One can mount a host path to the container via --volumes
or -v
. For example, to override the internal YAML config, one can do
# assuming $pwd is the root dir of this repo
docker run --rm -v $(pwd)/hub/example/mwu_encoder_ext.yml:/ext.yml jinaai/hub.examples.mwu_encoder --yaml-path /ext.yml
MWUEncoder@ 1[S]:look at me! im from an external yaml!
MWUEncoder@ 1[S]:initialize MWUEncoder from a yaml config
BasePea-0@ 1[I]:setting up sockets...
BasePea-0@ 1[I]:input tcp://0.0.0.0:36109 (PULL_BIND) output tcp://0.0.0.0:58191 (PUSH_BIND) control over tcp://0.0.0.0:52365 (PAIR_BIND)
BasePea-0@ 1[S]:ready and listening
To override the predefined entrypoint via --entrypoint
, e.g.
docker run --rm --entrypoint "jina" jinaai/hub.examples.mwu_encoder check
Another way to use the Pod image is simply giving it to jina pod
via --image
,
jina pod --image jinaai/hub.examples.mwu_encoder
π³ MWUEncoder@ 1[S]:look at me! im from internal yaml!
π³ MWUEncoder@ 1[S]:initialize MWUEncoder from a yaml config
π³ BasePea-0@ 1[I]:setting up sockets...
π³ BasePea-0@ 1[I]:input tcp://0.0.0.0:59608 (PULL_BIND) output tcp://0.0.0.0:59609 (PUSH_BIND) control over tcp://0.0.0.0:59610 (PAIR_BIND)
ContainerP@69041[S]:ready and listening
π³ BasePea-0@ 1[S]:ready and listening
Note the π³ represents that the log is piping from a Docker container.
See jina pod --help
for more usage.
Finally, one can use it via Flow API as well, e.g.
from jina.flow import Flow
f = (Flow()
.add(name='my-encoder', image='jinaai/hub.examples.mwu_encoder',
volumes='./abc', yaml_path='hub/examples/mwu-encoder/mwu_encoder_ext.yml',
port_in=55555, port_out=55556)
.add(name='my-indexer', yaml_path='indexer.yml'))
You can contribute your executor into this repository and it will be automatically built into an image and published to the Jina Hub via our CI/CD pipeline.
Typically, the following files are required:
File | Descriptions |
---|---|
Dockerfile |
describes the dependency setup and expose the entry point; |
manifest.yml |
metadata of the image, author, tags, etc. help the Hub to index and classify your image |
README.md |
an instruction for guiding users to use your image |
*.py |
describes the executor logic written in Python, if applicable; |
*.yml |
a YAML file describes the executor arguments and configs, if you want users to use your config; |
Note, large binary files (such as pretrained model, auxiliary data) are not recommended to upload to this repository. You can use RUN wget ...
or RUN curl
inside the Dockerfile
to download them from the web during the build.
Your file bundle awesomeness
should be uploaded to:
hub/
|
- executors/
|
- encoders/
|
- awesomeness/
|
|- Dockerfile
|- manifest.yml
|- README.md
|- awesome_ai.py
Your image will be published as jinaai/hub.executors.encoders.awesomeness
.
manifest.yml
must exist if you want to publish your Pod image to Jina Hub.
manifest.yml
annotates your image so that it can be managed by Jina Hub. To get better appealing on Jina Hub, you should carefully set manifest.yml
to the correct values.
Key | Description | Default |
---|---|---|
name |
Human-readable title of the image (must match with ^[a-zA-Z_$][a-zA-Z_\s\-$0-9]{3,20}$ ) |
Required |
description |
Human-readable description of the software packaged in the image | Required |
author |
Contact details of the people or organization responsible for the image (string) | Jina AI Dev-Team (dev-team@jina.ai) |
url |
URL to find more information on the image (string) | https://jina.ai |
documentation |
URL to get documentation on the image (string) | https://docs.jina.ai |
version |
Version of the image, it should be Semantic versioning-compatible | 0.0.0 |
vendor |
The name of the distributing entity, organization or individual (string) | Jina AI Limited |
license |
License under which contained software is distributed, it should be in this list | apache-2.0 |
avatar |
A picture that personalizes and distinguishes your image | None |
platform |
A list of CPU architectures that your image built on, each item should be in this list | [linux/amd64] |
Please refer to hub/examples/mwu_encoder/manifest.yml for the example.
All you need is to publish your bundle into this repo, the Docker image building, uploading and tagging are all handled automatically by our CICD pipeline.
- Let's say you organize all files mentioned in here in a folder called
awesomeness
. Depending on what you are contributing, you can put it intohub/executors/indexers/awesomeness
. - Make a Pull Request and commit your changes into this repository, remember to follow the commit lint style.
- Wait until the CICD finish and you get at least one reviewer's approval.
- Merge it!
The CICD pipeline will work on building, uploading and tagging the image on the Jina Hub.
The image will be available at jinaai/hub.executors.indexers.awesomeness:0.0.0
assuming your version number is defined as 0.0.0
in manifest.yml
.
You can use the image as the ways we described here.
Here is the checklist to help you locate the problem.
- Required file
Dockerfile
,manifest.yml
,README.md
is missing. - The required field in
manifest.yml
is missing. - Some field value is not in the correct format, not passing the sanity check.
- The pod bundle is badly placed.
- Time of building + testing is longer than 10 minutes.
- The build is successful but it fails on three basic usage tests.
Click "Details" and checkout the log of the CICD pipeline:
We welcome all kinds of contributions from the open-source community, individuals and partners. Without your active involvement, Jina won't be successful.
Please first read the contributing guidelines before the submission.
Copyright (c) 2020 Jina AI Limited. All rights reserved.
Jina is licensed under the Apache License, Version 2.0. See LICENSE for the full license text.