Dockerize your Executor#
Once you have understood what an Executor
is and how it can be used inside a Flow
, you may be interested in wrapping this Executor into a container
so that you can isolate its dependencies and make it ready to run in the cloud or in Kubernetes.
One option is to leverage Jina Hub infrastructure to make sure your Executor can run as a container.
However, you can also build a Docker image yourself and use it like any other Executor. There are some requirements on how this image needs to be built, the main ones being:
Jina must be installed inside the image
The Jina CLI command to start the Executor has to be the default entrypoint
Prerequisites#
To understand how a container image for an Executor is built, you need a basic understanding of Docker, both of how to write a Dockerfile, and how to build a Docker image.
To reproduce the example below it is required to have Docker installed locally.
Installing Jina in the Docker image#
Jina must be installed inside the Docker image. This can be achieved in one of two ways:
Use a Jina based image as the base image in your Dockerfile. This will make sure that everything needed for Jina to run the Executor is installed.
FROM jinaai/jina:3.0-py37-perf
Install Jina like any other Python package. You can do this by specifying Jina in the
requirements.txt
, or by including thepip install jina
command as part of the image building process.
RUN pip install jina==3.0
Setting Jina Executor CLI as entrypoint#
When a containerized Executor is run inside a Flow,
under the hood Jina executes docker run
with extra arguments.
This means that Jina assumes that whatever runs inside the container, also runs like it would in a regular OS process. Therefore, you need to make sure that
the basic entrypoint of the image calls jina executor
CLI command.
ENTRYPOINT ["jina", "executor", "--uses", "PATH_TO_YOUR_EXECUTOR_CONFIGURATION"]
Example: Dockerized Executor#
Here we will show how to build a basic Executor with a dependency on another external package
Writing the Executor#
You can define your soon-to-be-Dockerized Executor exactly like any other Executor.
We do this here in the my_executor.py
file.
import torch # Our Executor has dependency on torch
from jina import Executor, requests
class ContainerizedEncoder(Executor):
@requests
def foo(self, docs, **kwargs):
for doc in docs:
doc.text = 'This Document is embedded by ContainerizedEncoder'
doc.embedding = torch.randn(10)
Writing the Executor YAML file#
The YAML configuration, as a minimal working example, is required to point to the file containing the Executor.
More YAML options
To discover what else can be configured using Jina’s YAML interface, see here.
This is necessary for the Executor to be put inside the Docker image,
and we can define such a configuration in config.yml
:
jtype: ContainerizedEncoder
metas:
py_modules:
- my_executor.py
Writing requirements.txt
#
In this case, our Executor has only one requirement besides Jina: torch
.
In requirements.txt
, we specify a single requirement:
torch
Writing the Dockerfile#
The last step is to write a Dockerfile
, which has to do little more than launching the Executor via the Jina CLI:
FROM jinaai/jina:3.0-py37-perf
# make sure the files are copied into the image
COPY . /executor_root/
WORKDIR /executor_root
RUN pip install -r requirements.txt
ENTRYPOINT ["jina", "executor", "--uses", "config.yml"]
Building the image#
At this point we have a folder structure that looks like this:
.
├── my_executor.py
└── requirements.txt
└── config.yml
└── Dockerfile
We just need to build the image:
docker build -t my_containerized_executor .
Once the build is successful, this is what you should see under docker images
:
REPOSITORY TAG IMAGE ID CREATED SIZE
my_containerized_executor latest 5cead0161cb5 13 seconds ago 2.21GB
Using the containerized Executor#
The containerized Executor can be used like any other, the only difference being the ‘docker’ prefix in the uses
parameter:
from docarray import DocumentArray, Document
from jina import Flow
f = Flow().add(uses='docker://my_containerized_executor')
with f:
returned_docs = f.post(on='/', inputs=DocumentArray([Document()]))
for doc in returned_docs:
print(f'Document returned with text: "{doc.text}"')
print(f'Document embedding of shape {doc.embedding.shape}')
Document returned with text: "This Document is embedded by ContainerizedEncoder"
Document embedding of shape torch.Size([10])