Configuration
- name: <string>
kind: AsyncAPI
handler: # detailed configuration below
compute: # detailed configuration below
autoscaling: # detailed configuration below
update_strategy: # detailed configuration below
networking: # detailed configuration belowHandler
Python Handler
handler:
type: python
path: <string> # path to a python file with a Handler class definition, relative to the Cortex root (required)
dependencies: # (optional)
pip: <string> # relative path to requirements.txt (default: requirements.txt)
conda: <string> # relative path to conda-packages.txt (default: conda-packages.txt)
shell: <string> # relative path to a shell script for system package installation (default: dependencies.sh)
config: <string: value> # arbitrary dictionary passed to the constructor of the Handler class (optional)
python_path: <string> # path to the root of your Python folder that will be appended to PYTHONPATH (default: folder containing cortex.yaml)
image: <string> # docker image to use for the handler (default: quay.io/cortexlabs/python-handler-cpu:0.35.0, quay.io/cortexlabs/python-handler-gpu:0.35.0-cuda10.2-cudnn8, or quay.io/cortexlabs/python-handler-inf:0.35.0 based on compute)
env: <string: string> # dictionary of environment variables
log_level: <string> # log level that can be "debug", "info", "warning" or "error" (default: "info")
shm_size: <string> # size of shared memory (/dev/shm) for sharing data between multiple processes, e.g. 64Mi or 1Gi (default: Null)Tensorflow Handler
Compute
Autoscaling
Update strategy
Networking
Last updated