-
Notifications
You must be signed in to change notification settings - Fork 0
/
values.yaml
78 lines (65 loc) · 1.54 KB
/
values.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
# Default values for morpheus-ai-engine
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
replicaCount: 1
ngc:
username: "$oauthtoken"
apiKey: ""
org: ""
team: ""
aiengine:
registry: "nvcr.io/nvidia"
image: tritonserver
version: 22.12-py3
# don't use command due to nvidia_entrypoint.sh!
args:
- tritonserver
- --model-repository
- /common/triton-model-repo
- --model-control-mode
- explicit
# ensure triton gets at least one GPU
# and make it Pending the driver install
# assumes GPU device plugin
resources:
limits:
nvidia.com/gpu: 1
# use the following if we aren't using GPU device plugin
# resources: {}
broker:
registry: "docker.io"
image: bitnami/kafka
version: 2.7.0
# set brokerHost to a routable IP/hostname
# for connecting outside of host
brokerExternal:
brokerHost: localhost
brokerPort: 30092
zookeeper:
registry: "docker.io"
image: zookeeper
version: 3.6.3
hostCommonPath: /opt/morpheus/common
# use 'Always' for development
imagePullPolicy: IfNotPresent
imagePullSecrets:
- name: nvidia-registrykey-secret
# for Fleet Command deployments
# - name: imagepullsecret
# we need privileged pods to
# use hostPaths on OpenShift
serviceAccount:
create: false
name: morpheus
platform:
openshift: false
# deploy a LB (AWS) instead
# of a NodePort
loadBalancer:
enabled: false
# use a nodeSelector with
# OpenShift for GPU affinity
# default is nil for
# non GPU Operator/NFD use case
# see engine and sdk deployments
nodeSelector: {}