{{- if .Values.h2ogpt.enabled }} apiVersion: v1 kind: Service metadata: name: {{ include "h2ogpt.fullname" . }}-web namespace: {{ .Release.Namespace }} {{- with .Values.h2ogpt.service.webServiceAnnotations }} annotations: {{- toYaml . | nindent 4 }} {{- end }} spec: selector: app: {{ include "h2ogpt.fullname" . }} ports: - protocol: TCP port: {{ .Values.h2ogpt.service.webPort }} targetPort: 7860 type: {{ .Values.h2ogpt.service.type }} {{- end }} --- {{- if .Values.h2ogpt.enabled }} apiVersion: v1 kind: Service metadata: name: {{ include "h2ogpt.fullname" . }} namespace: {{ .Release.Namespace }} spec: selector: app: {{ include "h2ogpt.fullname" . }} ports: - protocol: TCP port: {{ .Values.h2ogpt.service.gptPort }} targetPort: 8888 type: {{ .Values.h2ogpt.service.type }} {{- end }} --- {{- if and (.Values.tgi.enabled) (not .Values.h2ogpt.stack.enabled ) }} apiVersion: v1 kind: Service metadata: name: {{ include "h2ogpt.fullname" . }}-tgi-inference namespace: {{ .Release.Namespace }} spec: selector: app: {{ include "h2ogpt.fullname" . }}-tgi-inference ports: - protocol: TCP port: {{ .Values.tgi.service.port }} targetPort: 80 type: {{ .Values.tgi.service.type }} {{- end }} --- {{- if and (.Values.vllm.enabled) (not .Values.h2ogpt.stack.enabled ) }} apiVersion: v1 kind: Service metadata: name: {{ include "h2ogpt.fullname" . }}-vllm-inference namespace: {{ .Release.Namespace }} spec: selector: app: {{ include "h2ogpt.fullname" . }}-vllm-inference ports: - protocol: TCP port: {{ .Values.vllm.service.port }} targetPort: 5000 type: {{ .Values.vllm.service.type }} {{- end }}