oc-k8s/opencloud/charts/hydra
2024-12-02 13:20:11 +01:00
..
charts hydra bootstraping 2024-12-02 13:20:11 +01:00
files hydra bootstraping 2024-12-02 13:20:11 +01:00
templates hydra bootstraping 2024-12-02 13:20:11 +01:00
.helmignore hydra bootstraping 2024-12-02 13:20:11 +01:00
Chart.lock hydra bootstraping 2024-12-02 13:20:11 +01:00
Chart.yaml hydra bootstraping 2024-12-02 13:20:11 +01:00
README.md hydra bootstraping 2024-12-02 13:20:11 +01:00
values.yaml hydra bootstraping 2024-12-02 13:20:11 +01:00

hydra

Version: 0.50.1 Type: application AppVersion: v2.2.0

A Helm chart for deploying ORY Hydra in Kubernetes

Homepage: https://www.ory.sh/

Maintainers

Name Email Url
ORY Team hi@ory.sh https://www.ory.sh/

Source Code

Requirements

Repository Name Version
file://../hydra-maester hydra-maester(hydra-maester) 0.50.1
file://../ory-commons ory(ory-commons) 0.1.0

Values

Key Type Default Description
affinity object {}
configmap.hashSumEnabled bool true switch to false to prevent checksum annotations being maintained and propogated to the pods
cronjob.janitor.affinity object {} Configure node affinity
cronjob.janitor.annotations object {} Set custom cron job level annotations
cronjob.janitor.automountServiceAccountToken bool true Set automounting of the SA token
cronjob.janitor.customArgs list [] Configure the arguments of the entrypoint, overriding the default value
cronjob.janitor.customCommand list [] Configure a custom entrypoint, overriding the default value
cronjob.janitor.extraContainers string "" If you want to add extra sidecar containers.
cronjob.janitor.extraEnv list [] Array of extra envs to be passed to the cronjob. This takes precedence over deployment variables. Kubernetes format is expected. Value is processed with Helm tpl - name: FOO value: BAR
cronjob.janitor.extraInitContainers string "" If you want to add extra init containers. These are processed before the migration init container.
cronjob.janitor.extraVolumeMounts list []
cronjob.janitor.extraVolumes list [] If you want to mount external volume
cronjob.janitor.labels object {} Set custom cron job level labels
cronjob.janitor.nodeSelector object {} Configure node labels for pod assignment
cronjob.janitor.podMetadata object {"annotations":{},"labels":{}} Specify pod metadata, this metadata is added directly to the pod, and not higher objects
cronjob.janitor.podMetadata.annotations object {} Extra pod level annotations
cronjob.janitor.podMetadata.labels object {} Extra pod level labels
cronjob.janitor.podSecurityContext object {}
cronjob.janitor.resources object {"limits":{},"requests":{}} We usually recommend not to specify default resources and to leave this as a conscious choice for the user. This also increases chances charts run on environments with little resources, such as Minikube. If you do want to specify resources, uncomment the following lines, adjust them as necessary, and remove the curly braces after 'resources:'. limits: cpu: 100m memory: 128Mi requests: cpu: 100m memory: 128Mi
cronjob.janitor.schedule string "0 */1 * * *" Configure how often the cron job is ran
cronjob.janitor.securityContext object {"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]},"privileged":false,"readOnlyRootFilesystem":true,"runAsNonRoot":true,"runAsUser":100} Configure the containers' SecurityContext for the janitor cronjob
cronjob.janitor.serviceAccount object {"annotations":{"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"0"},"create":true,"name":""} Specify the serviceAccountName value. In some situations it is needed to provides specific permissions to Hydra deployments Like for example installing Hydra on a cluster with a PosSecurityPolicy and Istio. Uncoment if it is needed to provide a ServiceAccount for the Hydra deployment.
cronjob.janitor.serviceAccount.annotations object {"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"0"} Annotations to add to the service account
cronjob.janitor.serviceAccount.create bool true Specifies whether a service account should be created
cronjob.janitor.serviceAccount.name string "" The name of the service account to use. If not set and create is true, a name is generated using the fullname template
cronjob.janitor.tolerations list [] Configure node tolerations
deployment.annotations object {} Set custom deployment level annotations
deployment.automigration object {"extraEnv":[]} Parameters for the automigration initContainer
deployment.automigration.extraEnv list [] Array of extra envs to be passed to the initContainer. Kubernetes format is expected. Value is processed with Helm tpl - name: FOO value: BAR
deployment.automountServiceAccountToken bool false
deployment.autoscaling object {"behavior":{},"enabled":false,"maxReplicas":3,"minReplicas":1,"targetCPU":{},"targetMemory":{}} Configure HPA
deployment.autoscaling.behavior object {} Set custom behavior https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/#configurable-scaling-behavior
deployment.customLivenessProbe object {} Configure a custom livenessProbe. This overwrites the default object
deployment.customReadinessProbe object {} Configure a custom readinessProbe. This overwrites the default object
deployment.customStartupProbe object {} Configure a custom startupProbe. This overwrites the default object
deployment.dnsConfig object {} Configure pod dnsConfig.
deployment.extraContainers string "" If you want to add extra sidecar containers.
deployment.extraEnv list [] Array of extra envs to be passed to the deployment. Kubernetes format is expected. Value is processed with Helm tpl - name: FOO value: BAR
deployment.extraInitContainers string "" If you want to add extra init containers. These are processed before the migration init container.
deployment.extraVolumeMounts list []
deployment.extraVolumes list [] If you want to mount external volume
deployment.initContainerSecurityContext object {}
deployment.labels object {} Set custom deployment level labels
deployment.lifecycle object {}
deployment.nodeSelector object {} Node labels for pod assignment.
deployment.podMetadata object {"annotations":{},"labels":{}} Specify pod metadata, this metadata is added directly to the pod, and not higher objects
deployment.podMetadata.annotations object {} Extra pod level annotations
deployment.podMetadata.labels object {} Extra pod level labels
deployment.podSecurityContext.fsGroup int 65534
deployment.podSecurityContext.fsGroupChangePolicy string "OnRootMismatch"
deployment.podSecurityContext.runAsGroup int 65534
deployment.podSecurityContext.runAsNonRoot bool true
deployment.podSecurityContext.runAsUser int 65534
deployment.podSecurityContext.seccompProfile.type string "RuntimeDefault"
deployment.readinessProbe object {"failureThreshold":5,"initialDelaySeconds":5,"periodSeconds":10} Default probe timers
deployment.resources object {} We usually recommend not to specify default resources and to leave this as a conscious choice for the user. This also increases chances charts run on environments with little resources, such as Minikube. If you do want to specify resources, uncomment the following lines, adjust them as necessary, and remove the curly braces after 'resources:'. limits: cpu: 100m memory: 128Mi requests: cpu: 100m memory: 128Mi
deployment.revisionHistoryLimit int 5 Number of revisions kept in history
deployment.securityContext.allowPrivilegeEscalation bool false
deployment.securityContext.capabilities.drop[0] string "ALL"
deployment.securityContext.privileged bool false
deployment.securityContext.readOnlyRootFilesystem bool true
deployment.securityContext.runAsGroup int 65534
deployment.securityContext.runAsNonRoot bool true
deployment.securityContext.runAsUser int 65534
deployment.securityContext.seLinuxOptions.level string "s0:c123,c456"
deployment.securityContext.seccompProfile.type string "RuntimeDefault"
deployment.serviceAccount object {"annotations":{},"create":true,"name":""} Specify the serviceAccountName value. In some situations it is needed to provides specific permissions to Hydra deployments Like for example installing Hydra on a cluster with a PosSecurityPolicy and Istio. Uncoment if it is needed to provide a ServiceAccount for the Hydra deployment.
deployment.serviceAccount.annotations object {} Annotations to add to the service account
deployment.serviceAccount.create bool true Specifies whether a service account should be created
deployment.serviceAccount.name string "" The name of the service account to use. If not set and create is true, a name is generated using the fullname template
deployment.startupProbe object {"failureThreshold":5,"initialDelaySeconds":0,"periodSeconds":1,"successThreshold":1,"timeoutSeconds":1} Default probe timers
deployment.strategy.rollingUpdate.maxSurge string "25%"
deployment.strategy.rollingUpdate.maxUnavailable string "25%"
deployment.strategy.type string "RollingUpdate"
deployment.terminationGracePeriodSeconds int 60
deployment.tolerations list [] Configure node tolerations.
deployment.topologySpreadConstraints list [] Configure pod topologySpreadConstraints.
fullnameOverride string "" Full chart name override
hydra-maester.adminService.name string "" The service name value may need to be set if you use fullnameOverride for the parent chart
hydra.automigration.customArgs list [] Ability to override arguments of the entrypoint. Can be used in-depended of customCommand eg: - sleep 5; - kratos
hydra.automigration.customCommand list [] Ability to override the entrypoint of the automigration container (e.g. to source dynamic secrets or export environment dynamic variables)
hydra.automigration.enabled bool false
hydra.automigration.resources object {} resource requests and limits for the automigration initcontainer
hydra.automigration.type string "job" Configure the way to execute database migration. Possible values: job, initContainer When set to job, the migration will be executed as a job on release or upgrade. When set to initContainer, the migration will be executed when kratos pod is created Defaults to job
hydra.command list ["hydra"] Ability to override the entrypoint of hydra container (e.g. to source dynamic secrets or export environment dynamic variables)
hydra.config object {"secrets":{},"serve":{"admin":{"port":4445},"public":{"port":4444},"tls":{"allow_termination_from":["10.0.0.0/8","172.16.0.0/12","192.168.0.0/16"]}},"urls":{"self":{}}} The ORY Hydra configuration. For a full list of available settings, check: https://www.ory.sh/docs/hydra/reference/configuration
hydra.config.secrets object {} The secrets have to be provided as a string slice, example: system: - "OG5XbmxXa3dYeGplQXpQanYxeEFuRUFa" - "foo bar 123 456 lorem" - "foo bar 123 456 lorem 1" - "foo bar 123 456 lorem 2" - "foo bar 123 456 lorem 3"
hydra.config.urls object {"self":{}} Configure the urls used by hydra itself, such as the issuer. Note: some values are required for hydra to start, please refer to https://www.ory.sh/docs/hydra/self-hosted/kubernetes-helm-chart self: issuer: "https://public.hydra.localhost:4444/"
hydra.customArgs list [] Ability to override arguments of the entrypoint. Can be used in-depended of customCommand
hydra.dev bool false Enable dev mode, not secure in production environments
image.pullPolicy string "IfNotPresent" Image pull policy
image.repository string "oryd/hydra" ORY Hydra image
image.tag string "v2.2.0" ORY Hydra version
imagePullSecrets list [] Image pull secrets
ingress.admin.annotations object {}
ingress.admin.className string ""
ingress.admin.enabled bool false En-/Disable the api ingress.
ingress.admin.hosts[0].host string "admin.hydra.localhost"
ingress.admin.hosts[0].paths[0].path string "/"
ingress.admin.hosts[0].paths[0].pathType string "ImplementationSpecific"
ingress.public object {"annotations":{},"className":"","enabled":false,"hosts":[{"host":"public.hydra.localhost","paths":[{"path":"/","pathType":"ImplementationSpecific"}]}]} Configure ingress for the proxy port.
ingress.public.enabled bool false En-/Disable the proxy ingress.
janitor.batchSize int 100 Configure how many records are deleted with each iteration
janitor.cleanupGrants bool false Configure if the trust relationships must be cleaned up
janitor.cleanupRequests bool false Configure if the consent and authentication requests must be cleaned up
janitor.cleanupTokens bool false Configure if the access and refresh tokens must be cleaned up
janitor.enabled bool false Enable cleanup of stale database rows by periodically running the janitor command
janitor.limit int 10000 Configure how many records are retrieved from database for deletion
job.annotations object {"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"1"} If you do want to specify annotations, uncomment the following lines, adjust them as necessary, and remove the curly braces after 'annotations:'.
job.automountServiceAccountToken bool true Set automounting of the SA token
job.extraContainers string "" If you want to add extra sidecar containers.
job.extraEnv list [] Array of extra envs to be passed to the job. This takes precedence over deployment variables. Kubernetes format is expected. Value is processed with Helm tpl - name: FOO value: BAR
job.extraInitContainers string "" If you want to add extra init containers. extraInitContainers:
job.labels object {} Set custom deployment level labels
job.lifecycle string "" If you want to add lifecycle hooks.
job.nodeSelector object {} Node labels for pod assignment.
job.podMetadata object {"annotations":{},"labels":{}} Specify pod metadata, this metadata is added directly to the pod, and not higher objects
job.podMetadata.annotations object {} Extra pod level annotations
job.podMetadata.labels object {} Extra pod level labels
job.resources object {} resource requests and limits for the automigration job
job.serviceAccount object {"annotations":{"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"0"},"create":true,"name":""} Specify the serviceAccountName value. In some situations it is needed to provides specific permissions to Hydra deployments Like for example installing Hydra on a cluster with a PosSecurityPolicy and Istio. Uncoment if it is needed to provide a ServiceAccount for the Hydra deployment.
job.serviceAccount.annotations object {"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"0"} Annotations to add to the service account
job.serviceAccount.create bool true Specifies whether a service account should be created
job.serviceAccount.name string "" The name of the service account to use. If not set and create is true, a name is generated using the fullname template
job.shareProcessNamespace bool false Set sharing process namespace
job.spec.backoffLimit int 10 Set job back off limit
job.tolerations list [] Configure node tolerations.
maester.enabled bool true
nameOverride string ""
pdb.enabled bool false
pdb.spec.maxUnavailable string ""
pdb.spec.minAvailable string ""
priorityClassName string "" Pod priority https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/
replicaCount int 1 Number of ORY Hydra members
secret.enabled bool true switch to false to prevent creating the secret
secret.hashSumEnabled bool true switch to false to prevent checksum annotations being maintained and propogated to the pods
secret.nameOverride string "" Provide custom name of existing secret, or custom name of secret to be created
secret.secretAnnotations object {"helm.sh/hook":"pre-install, pre-upgrade","helm.sh/hook-delete-policy":"before-hook-creation","helm.sh/hook-weight":"0","helm.sh/resource-policy":"keep"} Annotations to be added to secret. Annotations are added only when secret is being created. Existing secret will not be modified.
service.admin object {"annotations":{},"enabled":true,"labels":{},"loadBalancerIP":"","metricsPath":"/admin/metrics/prometheus","name":"http","port":4445,"type":"ClusterIP"} Configures the Kubernetes service for the api port.
service.admin.annotations object {} If you do want to specify annotations, uncomment the following lines, adjust them as necessary, and remove the curly braces after 'annotations:'.
service.admin.enabled bool true En-/disable the service
service.admin.loadBalancerIP string "" The load balancer IP
service.admin.metricsPath string "/admin/metrics/prometheus" Path to the metrics endpoint
service.admin.name string "http" The service port name. Useful to set a custom service port name if it must follow a scheme (e.g. Istio)
service.admin.port int 4445 The service port
service.admin.type string "ClusterIP" The service type
service.public object {"annotations":{},"enabled":true,"labels":{},"loadBalancerIP":"","name":"http","port":4444,"type":"ClusterIP"} Configures the Kubernetes service for the proxy port.
service.public.annotations object {} If you do want to specify annotations, uncomment the following lines, adjust them as necessary, and remove the curly braces after 'annotations:'.
service.public.enabled bool true En-/disable the service
service.public.loadBalancerIP string "" The load balancer IP
service.public.name string "http" The service port name. Useful to set a custom service port name if it must follow a scheme (e.g. Istio)
service.public.port int 4444 The service port
service.public.type string "ClusterIP" The service type
serviceMonitor.enabled bool false switch to true to enable creating the ServiceMonitor
serviceMonitor.labels object {} Provide additionnal labels to the ServiceMonitor ressource metadata
serviceMonitor.scheme string "http" HTTP scheme to use for scraping.
serviceMonitor.scrapeInterval string "60s" Interval at which metrics should be scraped
serviceMonitor.scrapeTimeout string "30s" Timeout after which the scrape is ended
serviceMonitor.tlsConfig object {} TLS configuration to use when scraping the endpoint
test.busybox object {"repository":"busybox","tag":1} use a busybox image from another repository
test.labels object {} Provide additional labels to the test pod
watcher.automountServiceAccountToken bool true
watcher.enabled bool false
watcher.image string "oryd/k8s-toolbox:v0.0.7"
watcher.mountFile string "" Path to mounted file, which wil be monitored for changes. eg: /etc/secrets/my-secret/foo
watcher.podMetadata object {"annotations":{},"labels":{}} Specify pod metadata, this metadata is added directly to the pod, and not higher objects
watcher.podMetadata.annotations object {} Extra pod level annotations
watcher.podMetadata.labels object {} Extra pod level labels
watcher.podSecurityContext object {} pod securityContext for watcher deployment
watcher.resources object {}
watcher.revisionHistoryLimit int 5 Number of revisions kept in history
watcher.securityContext object {"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]},"privileged":false,"readOnlyRootFilesystem":true,"runAsNonRoot":true,"runAsUser":100,"seccompProfile":{"type":"RuntimeDefault"}} container securityContext for watcher deployment
watcher.watchLabelKey string "ory.sh/watcher" Label key used for managing applications

Autogenerated from chart metadata using helm-docs v1.14.2