From b721ed1822953c94b7c2e204895ac24622de9edf Mon Sep 17 00:00:00 2001 From: Ricardo Ribeiro Date: Fri, 29 Sep 2023 13:19:20 +0000 Subject: [PATCH] update: source github.com/jsonnet-libs/k8s@15725033 --- 0.13/_custom/endpoint.libsonnet | 252 + 0.13/_custom/pod_metrics_endpoint.libsonnet | 243 + 0.13/_custom/relabel_config.libsonnet | 40 + 0.13/_gen/monitoring/main.libsonnet | 6 + .../_gen/monitoring/v1/alertmanager.libsonnet | 2811 +++ 0.13/_gen/monitoring/v1/main.libsonnet | 11 + 0.13/_gen/monitoring/v1/podMonitor.libsonnet | 333 + 0.13/_gen/monitoring/v1/probe.libsonnet | 346 + 0.13/_gen/monitoring/v1/prometheus.libsonnet | 3799 ++++ .../monitoring/v1/prometheusRule.libsonnet | 95 + .../monitoring/v1/serviceMonitor.libsonnet | 345 + 0.13/_gen/monitoring/v1/thanosRuler.libsonnet | 2599 +++ .../v1alpha1/alertmanagerConfig.libsonnet | 2316 +++ 0.13/_gen/monitoring/v1alpha1/main.libsonnet | 7 + .../v1alpha1/prometheusAgent.libsonnet | 3200 ++++ .../v1alpha1/scrapeConfig.libsonnet | 532 + 0.13/gen.libsonnet | 5 + 0.13/main.libsonnet | 1 + docs/0.13/README.md | 13 + docs/0.13/monitoring/index.md | 10 + docs/0.13/monitoring/v1/alertmanager.md | 10573 ++++++++++++ docs/0.13/monitoring/v1/endpoint.md | 629 + docs/0.13/monitoring/v1/index.md | 18 + docs/0.13/monitoring/v1/podMetricsEndpoint.md | 602 + docs/0.13/monitoring/v1/podMonitor.md | 1266 ++ docs/0.13/monitoring/v1/probe.md | 1301 ++ docs/0.13/monitoring/v1/prometheus.md | 14325 ++++++++++++++++ docs/0.13/monitoring/v1/prometheusRule.md | 381 + docs/0.13/monitoring/v1/relabelConfig.md | 75 + docs/0.13/monitoring/v1/serviceMonitor.md | 1322 ++ docs/0.13/monitoring/v1/thanosRuler.md | 9855 +++++++++++ .../monitoring/v1alpha1/alertmanagerConfig.md | 8335 +++++++++ docs/0.13/monitoring/v1alpha1/endpoint.md | 629 + docs/0.13/monitoring/v1alpha1/index.md | 14 + .../monitoring/v1alpha1/podMetricsEndpoint.md | 602 + .../monitoring/v1alpha1/prometheusAgent.md | 12112 +++++++++++++ .../0.13/monitoring/v1alpha1/relabelConfig.md | 75 + docs/0.13/monitoring/v1alpha1/scrapeConfig.md | 1957 +++ docs/README.md | 1 + 39 files changed, 81036 insertions(+) create mode 100644 0.13/_custom/endpoint.libsonnet create mode 100644 0.13/_custom/pod_metrics_endpoint.libsonnet create mode 100644 0.13/_custom/relabel_config.libsonnet create mode 100644 0.13/_gen/monitoring/main.libsonnet create mode 100644 0.13/_gen/monitoring/v1/alertmanager.libsonnet create mode 100644 0.13/_gen/monitoring/v1/main.libsonnet create mode 100644 0.13/_gen/monitoring/v1/podMonitor.libsonnet create mode 100644 0.13/_gen/monitoring/v1/probe.libsonnet create mode 100644 0.13/_gen/monitoring/v1/prometheus.libsonnet create mode 100644 0.13/_gen/monitoring/v1/prometheusRule.libsonnet create mode 100644 0.13/_gen/monitoring/v1/serviceMonitor.libsonnet create mode 100644 0.13/_gen/monitoring/v1/thanosRuler.libsonnet create mode 100644 0.13/_gen/monitoring/v1alpha1/alertmanagerConfig.libsonnet create mode 100644 0.13/_gen/monitoring/v1alpha1/main.libsonnet create mode 100644 0.13/_gen/monitoring/v1alpha1/prometheusAgent.libsonnet create mode 100644 0.13/_gen/monitoring/v1alpha1/scrapeConfig.libsonnet create mode 100644 0.13/gen.libsonnet create mode 100644 0.13/main.libsonnet create mode 100644 docs/0.13/README.md create mode 100644 docs/0.13/monitoring/index.md create mode 100644 docs/0.13/monitoring/v1/alertmanager.md create mode 100644 docs/0.13/monitoring/v1/endpoint.md create mode 100644 docs/0.13/monitoring/v1/index.md create mode 100644 docs/0.13/monitoring/v1/podMetricsEndpoint.md create mode 100644 docs/0.13/monitoring/v1/podMonitor.md create mode 100644 docs/0.13/monitoring/v1/probe.md create mode 100644 docs/0.13/monitoring/v1/prometheus.md create mode 100644 docs/0.13/monitoring/v1/prometheusRule.md create mode 100644 docs/0.13/monitoring/v1/relabelConfig.md create mode 100644 docs/0.13/monitoring/v1/serviceMonitor.md create mode 100644 docs/0.13/monitoring/v1/thanosRuler.md create mode 100644 docs/0.13/monitoring/v1alpha1/alertmanagerConfig.md create mode 100644 docs/0.13/monitoring/v1alpha1/endpoint.md create mode 100644 docs/0.13/monitoring/v1alpha1/index.md create mode 100644 docs/0.13/monitoring/v1alpha1/podMetricsEndpoint.md create mode 100644 docs/0.13/monitoring/v1alpha1/prometheusAgent.md create mode 100644 docs/0.13/monitoring/v1alpha1/relabelConfig.md create mode 100644 docs/0.13/monitoring/v1alpha1/scrapeConfig.md diff --git a/0.13/_custom/endpoint.libsonnet b/0.13/_custom/endpoint.libsonnet new file mode 100644 index 0000000..53f4d2d --- /dev/null +++ b/0.13/_custom/endpoint.libsonnet @@ -0,0 +1,252 @@ +local d = import 'doc-util/main.libsonnet'; + +local endpoint = { + '#':: d.pkg(name='endpoint', url='', help='Endpoint is a helper to create endpoint configurations in ServiceMonitors. An endpoint defines a scrapeable endpoint serving Prometheus metrics.'), + + '#authorization':: d.obj(help='Authorization section for this endpoint'), + authorization: { + + '#withType':: d.fn('Set the authentication type. Defaults to Bearer, Basic will cause an error', [d.arg('type', d.T.string)]), + withType(type): { authorization+: { type: type } }, + + '#credentials':: d.obj(help="The secret's key that contains the credentials of the request"), + credentials: { + + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { authorization+: { credentials+: { optional: is_optional } } }, + }, + }, + + '#withPort':: d.fn('Name of the service port this endpoint refers to. Mutually exclusive with withTargetPort.', [d.arg('port', d.T.string)]), + withPort(port): { port: port }, + + '#withTargetPort':: d.fn('Name or number of the target port of the Pod behind the Service, the port must be specified with container port property. Mutually exclusive with withPort.', [d.arg('port', d.T.int)]), + withTargetPort(port): { targetPort: port }, + + '#withPath':: d.fn('HTTP path to scrape for metrics.', [d.arg('path', d.T.string)]), + withPath(path): { path: path }, + + '#withScheme':: d.fn('HTTP scheme to use for scraping.', [d.arg('scheme', d.T.string)]), + withScheme(scheme): { scheme: scheme }, + + '#withParams':: d.fn('Optional HTTP URL parameters.', [d.arg('params', d.T.object)]), + withParams(params): { params: params }, + + '#withInterval':: d.fn('Interval at which metrics should be scraped.', [d.arg('interval', d.T.string)]), + withInterval(interval): { interval: interval }, + + '#withScrapeTimeout':: d.fn('Timeout after which the scrape is ended.', [d.arg('timeout', d.T.string)]), + withScrapeTimeout(timeout): { scrapeTimeout: timeout }, + + '#tlsConfig':: d.obj('TLS configuration to use when scraping the endpoint'), + tlsConfig: { + '#ca':: d.obj('Struct containing the CA cert to use for the targets.'), + ca: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { ca+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { ca+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#cert':: d.obj('Struct containing the client cert file for the targets.'), + cert: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { cert+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { cert+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#keySecret':: d.obj('Secret containing the client key file for the targets.'), + keySecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { keySecret+: { optional: is_optional } } }, + }, + + '#withServerName':: d.fn('Used to verify the hostname for the targets.', [d.arg('name', d.T.string)]), + withServerName(name): { tlsConfig+: { serverName: name } }, + + '#withInsecureSkipVerify':: d.fn('Disable target certificate validation.', [d.arg('skip_verify', d.T.bool)]), + withInsecureSkipVerify(skip_verify): { tlsConfig+: { insecureSkipVerify: skip_verify } }, + + '#withCaFile':: d.fn('Path to the CA cert in the Prometheus container to use for the targets.', [d.arg('path', d.T.string)]), + withCaFile(path): { tlsConfig+: { caFile: path } }, + + '#withCertFile':: d.fn('Path to the client cert file in the Prometheus container for the targets.', [d.arg('path', d.T.string)]), + withCertFile(path): { tlsConfig+: { certFile: path } }, + + '#withKeyFile':: d.fn('Path to the client key file in the Prometheus container for the targets.', [d.arg('path', d.T.string)]), + withKeyFile(path): { tlsConfig+: { keyFile: path } }, + }, + + '#withBearerTokenFile':: d.fn('File to read bearer token for scraping targets.', [d.arg('path', d.T.string)]), + withBearerTokenFile(path): { bearerTokenFile: path }, + + '#bearerTokenSecret':: d.obj('Secret to mount to read bearer token for scraping targets. The secret needs to be in the same namespace as the service monitor and accessible by the Prometheus Operator.'), + bearerTokenSecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { bearerTokenSecret+: { key: key } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { bearerTokenSecret+: { name: name } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { bearerTokenSecret+: { optional: is_optional } }, + }, + + '#withHonorLabels':: d.fn('HonorLabels chooses the metric’s labels on collisions with target labels.', [d.arg('honor_labels', d.T.bool)]), + withHonorLabels(honor_labels): { honorLabels: honor_labels }, + + '#withHonorTimestamps':: d.fn('HonorTimestamps controls whether Prometheus respects the timestamps present in scraped data.', [d.arg('honor_timestamps', d.T.bool)]), + withHonorTimestamps(honor_timestamps): { honorTimestamps: honor_timestamps }, + + '#basicAuth':: d.obj('BasicAuth allow an endpoint to authenticate over basic authentication More info: https://prometheus.io/docs/operating/configuration/#endpoints'), + basicAuth: { + '#username':: d.obj('The secret in the service monitor namespace that contains the username for authentication.'), + username: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { basicAuth+: { username+: { optional: is_optional } } }, + }, + + '#password':: d.obj('The secret in the service monitor namespace that contains the password for authentication.'), + password: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { basicAuth+: { password+: { optional: is_optional } } }, + }, + }, + + '#oauth2':: d.obj('OAuth2 for the URL. Only valid in Prometheus versions 2.27.0 and newer.'), + oauth2: { + '#clientId':: d.obj('The secret or configmap containing the OAuth2 client id.'), + clientId: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientId+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientId+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientId+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientId+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientId+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientId+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#clientSecret':: d.obj('The secret containing the OAuth2 client secret.'), + clientSecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientSecret+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientSecret+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientSecret+: { optional: is_optional } } }, + }, + + '#withTokenUrl':: d.fn('The URL to fetch the token from.', [d.arg('url', d.T.string)]), + withTokenUrl(url): { oauth2+: { tokenUrl: url } }, + + '#withScopes':: d.fn('OAuth2 scopes used for the token request.', [d.arg('scopes', d.T.array)]), + withScopes(scopes): { oauth2+: { scopes: scopes } }, + + '#withEndpointParams':: d.fn('Parameters to append to the token URL', [d.arg('params', d.T.object)]), + withEndpointParams(params): { oauth2+: { endpointParams: params } }, + }, + + '#withMetricRelabelings':: d.fn('MetricRelabelConfigs to apply to samples before ingestion.', [d.arg('relabelings', d.T.array)]), + withMetricRelabelings(relabelings): { metricRelabelings: relabelings }, + + '#withRelabelings':: d.fn('RelabelConfigs to apply to samples before scraping. Prometheus Operator automatically adds relabelings for a few standard Kubernetes fields and replaces original scrape job name with __tmp_prometheus_job_name. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config', [d.arg('relabelings', d.T.array)]), + withRelabelings(relabelings): { relabelings: relabelings }, + + '#withProxyUrl':: d.fn('ProxyURL eg http://proxyserver:2195 Directs scrapes to proxy through this endpoint.', [d.arg('url', d.T.string)]), + withProxyUrl(url): { proxyUrl: url }, + + '#mixin': 'ignore', + mixin: self, +}; + +{ + monitoring+:: { + v1+: { + endpoint: endpoint, + }, + v1alpha1+: { + endpoint: endpoint, + }, + }, +} diff --git a/0.13/_custom/pod_metrics_endpoint.libsonnet b/0.13/_custom/pod_metrics_endpoint.libsonnet new file mode 100644 index 0000000..c23305b --- /dev/null +++ b/0.13/_custom/pod_metrics_endpoint.libsonnet @@ -0,0 +1,243 @@ +local d = import 'doc-util/main.libsonnet'; + +local podMetricsEndpoint = { + '#':: d.pkg(name='podMetricsEndpoint', url='', help='PodMetricsEndpoint defines a scrapeable endpoint of a Kubernetes Pod serving Prometheus metrics.'), + + '#authorization':: d.obj(help='Authorization section for this endpoint'), + authorization: { + + '#withType':: d.fn('Set the authentication type. Defaults to Bearer, Basic will cause an error', [d.arg('type', d.T.string)]), + withType(type): { authorization+: { type: type } }, + + '#credentials':: d.obj(help="The secret's key that contains the credentials of the request"), + credentials: { + + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { authorization+: { credentials+: { optional: is_optional } } }, + }, + }, + + '#withPort':: d.fn('Name of the pod port this endpoint refers to. Mutually exclusive with targetPort.', [d.arg('port', d.T.string)]), + withPort(port): { port: port }, + + '#withTargetPort':: d.fn('Deprecated: Use ‘port’ instead.', [d.arg('port', d.T.int)]), + withTargetPort(port): { targetPort: port }, + + '#withPath':: d.fn('HTTP path to scrape for metrics.', [d.arg('path', d.T.string)]), + withPath(path): { path: path }, + + '#withScheme':: d.fn('HTTP scheme to use for scraping.', [d.arg('scheme', d.T.string)]), + withScheme(scheme): { scheme: scheme }, + + '#withParams':: d.fn('Optional HTTP URL parameters.', [d.arg('params', d.T.object)]), + withParams(params): { params: params }, + + '#withInterval':: d.fn('Interval at which metrics should be scraped.', [d.arg('interval', d.T.string)]), + withInterval(interval): { interval: interval }, + + '#withScrapeTimeout':: d.fn('Timeout after which the scrape is ended.', [d.arg('timeout', d.T.string)]), + withScrapeTimeout(timeout): { scrapeTimeout: timeout }, + + '#tlsConfig':: d.obj('TLS configuration to use when scraping the endpoint'), + tlsConfig: { + '#ca':: d.obj('Struct containing the CA cert to use for the targets.'), + ca: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { ca+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { ca+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#cert':: d.obj('Struct containing the client cert file for the targets.'), + cert: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { cert+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { cert+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#keySecret':: d.obj('Secret containing the client key file for the targets.'), + keySecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { tlsConfig+: { keySecret+: { optional: is_optional } } }, + }, + + '#withServerName':: d.fn('Used to verify the hostname for the targets.', [d.arg('name', d.T.string)]), + withServerName(name): { tlsConfig+: { serverName: name } }, + + '#withInsecureSkipVerify':: d.fn('Disable target certificate validation.', [d.arg('skip_verify', d.T.bool)]), + withInsecureSkipVerify(skip_verify): { tlsConfig+: { insecureSkipVerify: skip_verify } }, + }, + + '#withBearerTokenFile':: d.fn('File to read bearer token for scraping targets.', [d.arg('path', d.T.string)]), + withBearerTokenFile(path): { bearerTokenFile: path }, + + '#bearerTokenSecret':: d.obj('Secret to mount to read bearer token for scraping targets. The secret needs to be in the same namespace as the service monitor and accessible by the Prometheus Operator.'), + bearerTokenSecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { bearerTokenSecret+: { key: key } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { bearerTokenSecret+: { name: name } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { bearerTokenSecret+: { optional: is_optional } }, + }, + + '#withHonorLabels':: d.fn('HonorLabels chooses the metric’s labels on collisions with target labels.', [d.arg('honor_labels', d.T.bool)]), + withHonorLabels(honor_labels): { honorLabels: honor_labels }, + + '#withHonorTimestamps':: d.fn('HonorTimestamps controls whether Prometheus respects the timestamps present in scraped data.', [d.arg('honor_timestamps', d.T.bool)]), + withHonorTimestamps(honor_timestamps): { honorTimestamps: honor_timestamps }, + + '#basicAuth':: d.obj('BasicAuth allow an endpoint to authenticate over basic authentication More info: https://prometheus.io/docs/operating/configuration/#endpoints'), + basicAuth: { + '#username':: d.obj('The secret in the service monitor namespace that contains the username for authentication.'), + username: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { basicAuth+: { username+: { optional: is_optional } } }, + }, + + '#password':: d.obj('The secret in the service monitor namespace that contains the password for authentication.'), + password: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { basicAuth+: { password+: { optional: is_optional } } }, + }, + }, + + '#oauth2':: d.obj('OAuth2 for the URL. Only valid in Prometheus versions 2.27.0 and newer.'), + oauth2: { + '#clientId':: d.obj('The secret or configmap containing the OAuth2 client id.'), + clientId: { + '#secret':: d.obj('Secret containing data to use for the targets. Mutually exclusive with configMap.'), + secret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientId+: { secret+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientId+: { secret+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientId+: { secret+: { optional: is_optional } } } }, + }, + + '#configMap':: d.obj('ConfigMap containing data to use for the targets. Mutually exclusive with secret.'), + configMap: { + '#withKey':: d.fn('The key to select.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientId+: { configMap+: { key: key } } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientId+: { configMap+: { name: name } } } }, + + '#withOptional':: d.fn('Specify whether the ConfigMap or its key must be defined.', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientId+: { configMap+: { optional: is_optional } } } }, + }, + }, + + '#clientSecret':: d.obj('The secret containing the OAuth2 client secret.'), + clientSecret: { + '#withKey':: d.fn('The key of the secret to select from. Must be a valid secret key.', [d.arg('key', d.T.string)]), + withKey(key): { oauth2+: { clientSecret+: { key: key } } }, + + '#withName':: d.fn('Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names', [d.arg('name', d.T.string)]), + withName(name): { oauth2+: { clientSecret+: { name: name } } }, + + '#withOptional':: d.fn('Specify whether the Secret or its key must be defined', [d.arg('is_optional', d.T.bool)]), + withOptional(is_optional): { oauth2+: { clientSecret+: { optional: is_optional } } }, + }, + + '#withTokenUrl':: d.fn('The URL to fetch the token from.', [d.arg('url', d.T.string)]), + withTokenUrl(url): { oauth2+: { tokenUrl: url } }, + + '#withScopes':: d.fn('OAuth2 scopes used for the token request.', [d.arg('scopes', d.T.array)]), + withScopes(scopes): { oauth2+: { scopes: scopes } }, + + '#withEndpointParams':: d.fn('Parameters to append to the token URL', [d.arg('params', d.T.object)]), + withEndpointParams(params): { oauth2+: { endpointParams: params } }, + }, + + '#withMetricRelabelings':: d.fn('MetricRelabelConfigs to apply to samples before ingestion.', [d.arg('relabelings', d.T.array)]), + withMetricRelabelings(relabelings): { metricRelabelings: relabelings }, + + '#withRelabelings':: d.fn('RelabelConfigs to apply to samples before scraping. Prometheus Operator automatically adds relabelings for a few standard Kubernetes fields and replaces original scrape job name with __tmp_prometheus_job_name. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config', [d.arg('relabelings', d.T.array)]), + withRelabelings(relabelings): { relabelings: relabelings }, + + '#withProxyUrl':: d.fn('ProxyURL eg http://proxyserver:2195 Directs scrapes to proxy through this endpoint.', [d.arg('url', d.T.string)]), + withProxyUrl(url): { proxyUrl: url }, + + '#mixin': 'ignore', + mixin: self, +}; + +{ + monitoring+:: { + v1+: { + podMetricsEndpoint: podMetricsEndpoint, + }, + v1alpha1+: { + podMetricsEndpoint: podMetricsEndpoint, + }, + }, +} diff --git a/0.13/_custom/relabel_config.libsonnet b/0.13/_custom/relabel_config.libsonnet new file mode 100644 index 0000000..eb4d364 --- /dev/null +++ b/0.13/_custom/relabel_config.libsonnet @@ -0,0 +1,40 @@ +local d = import 'doc-util/main.libsonnet'; + +local relabelConfig = { + '#':: d.pkg(name='relabelConfig', url='', help='RelabelConfig is a help to build RelabelConfig objects for ServiceMonitors and PodMonitors'), + + '#withSourceLabels':: d.fn('The source labels select values from existing labels. Their content is concatenated using the configured separator and matched against the configured regular expression for the replace, keep, and drop actions.', [d.arg('labels', d.T.array)]), + withSourceLabels(labels): { sourceLabels: labels }, + + '#withSeparator':: d.fn('Separator placed between concatenated source label values. default is ‘;’.', [d.arg('separator', d.T.string)]), + withSeparator(separator): { separator: separator }, + + '#withTargetLabel':: d.fn('Label to which the resulting value is written in a replace action. It is mandatory for replace actions. Regex capture groups are available.', [d.arg('separator', d.T.string)]), + withTargetLabel(label): { targetLabel: label }, + + '#withRegex':: d.fn('Regular expression against which the extracted value is matched. Default is ‘(.*)’', [d.arg('separator', d.T.string)]), + withRegex(regex): { regex: regex }, + + '#withModulus':: d.fn('Modulus to take of the hash of the source label values.', [d.arg('modulus', d.T.number)]), + withModulus(modulus): { modulus: modulus }, + + '#withReplacement':: d.fn('Replacement value against which a regex replace is performed if the regular expression matches. Regex capture groups are available. Default is ‘$1’', [d.arg('separator', d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + + '#withAction':: d.fn('Action to perform based on regex matching. Default is ‘replace’', [d.arg('separator', d.T.string)]), + withAction(action): { action: action }, + + '#mixin': 'ignore', + mixin: self, +}; + +{ + monitoring+:: { + v1+: { + relabelConfig: relabelConfig, + }, + v1alpha1+: { + relabelConfig: relabelConfig, + }, + }, +} diff --git a/0.13/_gen/monitoring/main.libsonnet b/0.13/_gen/monitoring/main.libsonnet new file mode 100644 index 0000000..a25fafd --- /dev/null +++ b/0.13/_gen/monitoring/main.libsonnet @@ -0,0 +1,6 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='monitoring', url='', help=''), + v1: (import 'v1/main.libsonnet'), + v1alpha1: (import 'v1alpha1/main.libsonnet'), +} diff --git a/0.13/_gen/monitoring/v1/alertmanager.libsonnet b/0.13/_gen/monitoring/v1/alertmanager.libsonnet new file mode 100644 index 0000000..e213332 --- /dev/null +++ b/0.13/_gen/monitoring/v1/alertmanager.libsonnet @@ -0,0 +1,2811 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='alertmanager', url='', help='"Alertmanager describes an Alertmanager cluster."'), + '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Alertmanager', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'monitoring.coreos.com/v1', + kind: 'Alertmanager', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='"Specification of the desired behavior of the Alertmanager cluster. More info: https://github.com/kubernetes/community/blob/master/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'), + spec: { + '#affinity':: d.obj(help="\"If specified, the pod's scheduling constraints.\""), + affinity: { + '#nodeAffinity':: d.obj(help='"Describes node affinity scheduling rules for the pod."'), + nodeAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#preference':: d.obj(help='"A node selector term, associated with the corresponding weight."'), + preference: { + '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""), + matchExpressions: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""), + matchFields: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { preference+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { preference+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { preference+: { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { preference+: { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + }, + '#withWeight':: d.fn(help='"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#nodeSelectorTerms':: d.obj(help='"Required. A list of node selector terms. The terms are ORed."'), + nodeSelectorTerms: { + '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""), + matchExpressions: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""), + matchFields: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + }, + '#withNodeSelectorTerms':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAffinity':: d.obj(help='"Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s))."'), + podAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'), + podAffinityTerm: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } }, + }, + '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAntiAffinity':: d.obj(help='"Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s))."'), + podAntiAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'), + podAffinityTerm: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } }, + }, + '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + }, + '#alertmanagerConfigMatcherStrategy':: d.obj(help='"The AlertmanagerConfigMatcherStrategy defines how AlertmanagerConfig objects match the alerts. In the future more options may be added."'), + alertmanagerConfigMatcherStrategy: { + '#withType':: d.fn(help='"If set to `OnNamespace`, the operator injects a label matcher matching the namespace of the AlertmanagerConfig object for all its routes and inhibition rules. `None` will not add any additional matchers other than the ones specified in the AlertmanagerConfig. Default is `OnNamespace`."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { alertmanagerConfigMatcherStrategy+: { type: type } } }, + }, + '#alertmanagerConfigNamespaceSelector':: d.obj(help='"Namespaces to be selected for AlertmanagerConfig discovery. If nil, only check own namespace."'), + alertmanagerConfigNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { alertmanagerConfigNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { alertmanagerConfigNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { alertmanagerConfigNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { alertmanagerConfigNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#alertmanagerConfigSelector':: d.obj(help='"AlertmanagerConfigs to be selected for to merge and configure Alertmanager with."'), + alertmanagerConfigSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { alertmanagerConfigSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { alertmanagerConfigSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { alertmanagerConfigSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { alertmanagerConfigSelector+: { matchLabels+: matchLabels } } }, + }, + '#alertmanagerConfiguration':: d.obj(help='"EXPERIMENTAL: alertmanagerConfiguration specifies the configuration of Alertmanager. If defined, it takes precedence over the `configSecret` field. This field may change in future releases."'), + alertmanagerConfiguration: { + '#global':: d.obj(help='"Defines the global parameters of the Alertmanager configuration."'), + global: { + '#httpConfig':: d.obj(help='"HTTP client configuration."'), + httpConfig: { + '#authorization':: d.obj(help='"Authorization header configuration for the client. This is mutually exclusive with BasicAuth and is only available starting from Alertmanager v0.22+."'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { authorization+: { credentials+: { key: key } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { authorization+: { credentials+: { name: name } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { authorization+: { credentials+: { optional: optional } } } } } } }, + }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { authorization+: { type: type } } } } } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth for the client. This is mutually exclusive with Authorization. If both are defined, BasicAuth takes precedence."'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { password+: { key: key } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { password+: { name: name } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { password+: { optional: optional } } } } } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { username+: { key: key } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { username+: { name: name } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { basicAuth+: { username+: { optional: optional } } } } } } }, + }, + }, + '#bearerTokenSecret':: d.obj(help="\"The secret's key that contains the bearer token to be used by the client for authentication. The secret needs to be in the same namespace as the Alertmanager object and accessible by the Prometheus Operator.\""), + bearerTokenSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { bearerTokenSecret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { bearerTokenSecret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { bearerTokenSecret+: { optional: optional } } } } } }, + }, + '#oauth2':: d.obj(help='"OAuth2 client credentials used to fetch a token for the targets."'), + oauth2: { + '#clientId':: d.obj(help='"The secret or configmap containing the OAuth2 client id"'), + clientId: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { configMap+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { configMap+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { configMap+: { optional: optional } } } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { secret+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { secret+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientId+: { secret+: { optional: optional } } } } } } } }, + }, + }, + '#clientSecret':: d.obj(help='"The secret containing the OAuth2 client secret"'), + clientSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientSecret+: { key: key } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientSecret+: { name: name } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { clientSecret+: { optional: optional } } } } } } }, + }, + '#withEndpointParams':: d.fn(help='"Parameters to append to the token URL"', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParams(endpointParams): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { endpointParams: endpointParams } } } } } }, + '#withEndpointParamsMixin':: d.fn(help='"Parameters to append to the token URL"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParamsMixin(endpointParams): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { endpointParams+: endpointParams } } } } } }, + '#withScopes':: d.fn(help='"OAuth2 scopes used for the token request"', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } } } } } }, + '#withScopesMixin':: d.fn(help='"OAuth2 scopes used for the token request"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } } } } } }, + '#withTokenUrl':: d.fn(help='"The URL to fetch the token from"', args=[d.arg(name='tokenUrl', type=d.T.string)]), + withTokenUrl(tokenUrl): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { oauth2+: { tokenUrl: tokenUrl } } } } } }, + }, + '#tlsConfig':: d.obj(help='"TLS configuration for the client."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { configMap+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { configMap+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { configMap+: { optional: optional } } } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { secret+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { secret+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { ca+: { secret+: { optional: optional } } } } } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { secret+: { key: key } } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { secret+: { name: name } } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { keySecret+: { key: key } } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { keySecret+: { name: name } } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { keySecret+: { optional: optional } } } } } } }, + }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } } } } } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { tlsConfig+: { serverName: serverName } } } } } }, + }, + '#withFollowRedirects':: d.fn(help='"FollowRedirects specifies whether the client should follow HTTP 3xx redirects."', args=[d.arg(name='followRedirects', type=d.T.boolean)]), + withFollowRedirects(followRedirects): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { followRedirects: followRedirects } } } } }, + '#withProxyURL':: d.fn(help='"Optional proxy URL."', args=[d.arg(name='proxyURL', type=d.T.string)]), + withProxyURL(proxyURL): { spec+: { alertmanagerConfiguration+: { global+: { httpConfig+: { proxyURL: proxyURL } } } } }, + }, + '#opsGenieApiKey':: d.obj(help='"The default OpsGenie API Key."'), + opsGenieApiKey: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiKey+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiKey+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiKey+: { optional: optional } } } } }, + }, + '#opsGenieApiUrl':: d.obj(help='"The default OpsGenie API URL."'), + opsGenieApiUrl: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiUrl+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiUrl+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { opsGenieApiUrl+: { optional: optional } } } } }, + }, + '#slackApiUrl':: d.obj(help='"The default Slack API URL."'), + slackApiUrl: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { slackApiUrl+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { slackApiUrl+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { slackApiUrl+: { optional: optional } } } } }, + }, + '#smtp':: d.obj(help='"Configures global SMTP parameters."'), + smtp: { + '#authPassword':: d.obj(help='"SMTP Auth using LOGIN and PLAIN."'), + authPassword: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authPassword+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authPassword+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authPassword+: { optional: optional } } } } } }, + }, + '#authSecret':: d.obj(help='"SMTP Auth using CRAM-MD5."'), + authSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authSecret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authSecret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authSecret+: { optional: optional } } } } } }, + }, + '#smartHost':: d.obj(help='"The default SMTP smarthost used for sending emails."'), + smartHost: { + '#withHost':: d.fn(help="\"Defines the host's address, it can be a DNS name or a literal IP address.\"", args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { smartHost+: { host: host } } } } } }, + '#withPort':: d.fn(help="\"Defines the host's port, it can be a literal port number or a port name.\"", args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { smartHost+: { port: port } } } } } }, + }, + '#withAuthIdentity':: d.fn(help='"SMTP Auth using PLAIN"', args=[d.arg(name='authIdentity', type=d.T.string)]), + withAuthIdentity(authIdentity): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authIdentity: authIdentity } } } } }, + '#withAuthUsername':: d.fn(help="\"SMTP Auth using CRAM-MD5, LOGIN and PLAIN. If empty, Alertmanager doesn't authenticate to the SMTP server.\"", args=[d.arg(name='authUsername', type=d.T.string)]), + withAuthUsername(authUsername): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { authUsername: authUsername } } } } }, + '#withFrom':: d.fn(help='"The default SMTP From header field."', args=[d.arg(name='from', type=d.T.string)]), + withFrom(from): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { from: from } } } } }, + '#withHello':: d.fn(help='"The default hostname to identify to the SMTP server."', args=[d.arg(name='hello', type=d.T.string)]), + withHello(hello): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { hello: hello } } } } }, + '#withRequireTLS':: d.fn(help='"The default SMTP TLS requirement. Note that Go does not support unencrypted connections to remote SMTP endpoints."', args=[d.arg(name='requireTLS', type=d.T.boolean)]), + withRequireTLS(requireTLS): { spec+: { alertmanagerConfiguration+: { global+: { smtp+: { requireTLS: requireTLS } } } } }, + }, + '#withPagerdutyUrl':: d.fn(help='"The default Pagerduty URL."', args=[d.arg(name='pagerdutyUrl', type=d.T.string)]), + withPagerdutyUrl(pagerdutyUrl): { spec+: { alertmanagerConfiguration+: { global+: { pagerdutyUrl: pagerdutyUrl } } } }, + '#withResolveTimeout':: d.fn(help='"ResolveTimeout is the default value used by alertmanager if the alert does not include EndsAt, after this time passes it can declare the alert as resolved if it has not been updated. This has no impact on alerts from Prometheus, as they always include EndsAt."', args=[d.arg(name='resolveTimeout', type=d.T.string)]), + withResolveTimeout(resolveTimeout): { spec+: { alertmanagerConfiguration+: { global+: { resolveTimeout: resolveTimeout } } } }, + }, + '#templates':: d.obj(help='"Custom notification templates."'), + templates: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { configMap+: { key: key } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { secret+: { key: key } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secret+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + }, + }, + '#withName':: d.fn(help='"The name of the AlertmanagerConfig resource which is used to generate the Alertmanager configuration. It must be defined in the same namespace as the Alertmanager object. The operator will not enforce a `namespace` label for routes and inhibition rules."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { alertmanagerConfiguration+: { name: name } } }, + '#withTemplates':: d.fn(help='"Custom notification templates."', args=[d.arg(name='templates', type=d.T.array)]), + withTemplates(templates): { spec+: { alertmanagerConfiguration+: { templates: if std.isArray(v=templates) then templates else [templates] } } }, + '#withTemplatesMixin':: d.fn(help='"Custom notification templates."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='templates', type=d.T.array)]), + withTemplatesMixin(templates): { spec+: { alertmanagerConfiguration+: { templates+: if std.isArray(v=templates) then templates else [templates] } } }, + }, + '#containers':: d.obj(help='"Containers allows injecting additional containers. This is meant to allow adding an authentication proxy to an Alertmanager pod. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. The current container names are: `alertmanager` and `config-reloader`. Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."'), + containers: { + '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'), + env: { + '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""), + valueFrom: { + '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'), + configMapKeyRef: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } }, + }, + '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } }, + }, + '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""), + secretKeyRef: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { secretKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } }, + }, + }, + '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'), + envFrom: { + '#configMapRef':: d.obj(help='"The ConfigMap to select from"'), + configMapRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapRef+: { optional: optional } }, + }, + '#secretRef':: d.obj(help='"The Secret to select from"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretRef+: { optional: optional } }, + }, + '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]), + withPrefix(prefix): { prefix: prefix }, + }, + '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'), + lifecycle: { + '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'), + postStart: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""), + preStop: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + livenessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { livenessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { livenessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'), + ports: { + '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]), + withContainerPort(containerPort): { containerPort: containerPort }, + '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]), + withHostPort(hostPort): { hostPort: hostPort }, + '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + }, + '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + readinessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { readinessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { readinessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'), + resizePolicy: { + '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]), + withResourceName(resourceName): { resourceName: resourceName }, + '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy }, + }, + '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'), + securityContext: { + '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'), + capabilities: { + '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""), + startupProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { startupProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { startupProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'), + volumeDevices: { + '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + }, + '#hostAliases':: d.obj(help="\"Pods' hostAliases configuration\""), + hostAliases: { + '#withHostnames':: d.fn(help='"Hostnames for the above IP address."', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnames(hostnames): { hostnames: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withHostnamesMixin':: d.fn(help='"Hostnames for the above IP address."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnamesMixin(hostnames): { hostnames+: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withIp':: d.fn(help='"IP address of the host file entry."', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + }, + '#imagePullSecrets':: d.obj(help='"An optional list of references to secrets in the same namespace to use for pulling prometheus and alertmanager images from registries see http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"'), + imagePullSecrets: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#initContainers':: d.obj(help='"InitContainers allows adding initContainers to the pod definition. Those can be used to e.g. fetch secrets for injection into the Alertmanager configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. The current init container name is: `init-config-reloader`. Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."'), + initContainers: { + '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'), + env: { + '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""), + valueFrom: { + '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'), + configMapKeyRef: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } }, + }, + '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } }, + }, + '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""), + secretKeyRef: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { secretKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } }, + }, + }, + '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'), + envFrom: { + '#configMapRef':: d.obj(help='"The ConfigMap to select from"'), + configMapRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapRef+: { optional: optional } }, + }, + '#secretRef':: d.obj(help='"The Secret to select from"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretRef+: { optional: optional } }, + }, + '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]), + withPrefix(prefix): { prefix: prefix }, + }, + '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'), + lifecycle: { + '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'), + postStart: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""), + preStop: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + livenessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { livenessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { livenessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'), + ports: { + '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]), + withContainerPort(containerPort): { containerPort: containerPort }, + '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]), + withHostPort(hostPort): { hostPort: hostPort }, + '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + }, + '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + readinessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { readinessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { readinessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'), + resizePolicy: { + '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]), + withResourceName(resourceName): { resourceName: resourceName }, + '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy }, + }, + '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'), + securityContext: { + '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'), + capabilities: { + '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""), + startupProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { startupProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { startupProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'), + volumeDevices: { + '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + }, + '#podMetadata':: d.obj(help='"PodMetadata configures Labels and Annotations which are propagated to the alertmanager pods."'), + podMetadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { podMetadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { podMetadata+: { annotations+: annotations } } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { podMetadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { podMetadata+: { labels+: labels } } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { podMetadata+: { name: name } } }, + }, + '#resources':: d.obj(help='"Define resources requests and limits for single Pods."'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { resources+: { limits: limits } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { resources+: { limits+: limits } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { resources+: { requests: requests } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { resources+: { requests+: requests } } }, + }, + '#securityContext':: d.obj(help='"SecurityContext holds pod-level security attributes and common container settings. This defaults to the default PodSecurityContext."'), + securityContext: { + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by the containers in this pod. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seccompProfile+: { type: type } } } }, + }, + '#sysctls':: d.obj(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."'), + sysctls: { + '#withName':: d.fn(help='"Name of a property to set"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Value of a property to set"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#windowsOptions':: d.obj(help="\"The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux.\""), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { spec+: { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } }, + }, + '#withFsGroup':: d.fn(help="\"A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: \\n 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- \\n If unset, the Kubelet will not modify the ownership and permissions of any volume. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { securityContext+: { fsGroup: fsGroup } } }, + '#withFsGroupChangePolicy':: d.fn(help='"fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are \\"OnRootMismatch\\" and \\"Always\\". If not specified, \\"Always\\" is used. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { securityContext+: { runAsGroup: runAsGroup } } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { securityContext+: { runAsUser: runAsUser } } }, + '#withSupplementalGroups':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSupplementalGroupsMixin':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSysctls':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + '#withSysctlsMixin':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + }, + '#storage':: d.obj(help='"Storage is the definition of how storage will be used by the Alertmanager instances."'), + storage: { + '#emptyDir':: d.obj(help='"EmptyDirVolumeSource to be used by the StatefulSet. If specified, it takes precedence over `ephemeral` and `volumeClaimTemplate`. More info: https://kubernetes.io/docs/concepts/storage/volumes/#emptydir"'), + emptyDir: { + '#withMedium':: d.fn(help="\"medium represents what type of storage medium should back this directory. The default is \\\"\\\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\"", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { spec+: { storage+: { emptyDir+: { medium: medium } } } }, + '#withSizeLimit':: d.fn(help='"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"', args=[d.arg(name='sizeLimit', type=d.T.any)]), + withSizeLimit(sizeLimit): { spec+: { storage+: { emptyDir+: { sizeLimit: sizeLimit } } } }, + }, + '#ephemeral':: d.obj(help='"EphemeralVolumeSource to be used by the StatefulSet. This is a beta field in k8s 1.21 and GA in 1.15. For lower versions, starting with k8s 1.19, it requires enabling the GenericEphemeralVolume feature gate. More info: https://kubernetes.io/docs/concepts/storage/ephemeral-volumes/#generic-ephemeral-volumes"'), + ephemeral: { + '#volumeClaimTemplate':: d.obj(help='"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \\n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \\n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \\n Required, must not be nil."'), + volumeClaimTemplate: { + '#spec':: d.obj(help='"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } } } }, + }, + '#withMetadata':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadata(metadata): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { metadata: metadata } } } } }, + '#withMetadataMixin':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadataMixin(metadata): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { metadata+: metadata } } } } }, + }, + }, + '#volumeClaimTemplate':: d.obj(help='"Defines the PVC spec to be used by the Prometheus StatefulSets. The easiest way to use a volume that cannot be automatically provisioned is to use a label selector alongside manually created PersistentVolumes."'), + volumeClaimTemplate: { + '#metadata':: d.obj(help='"EmbeddedMetadata contains metadata relevant to an EmbeddedResource."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { annotations: annotations } } } } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { annotations+: annotations } } } } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { labels: labels } } } } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { labels+: labels } } } } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { name: name } } } } }, + }, + '#spec':: d.obj(help='"Defines the desired characteristics of a volume requested by a pod author. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } } }, + }, + '#withApiVersion':: d.fn(help='"APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources"', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { storage+: { volumeClaimTemplate+: { apiVersion: apiVersion } } } }, + '#withKind':: d.fn(help='"Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { kind: kind } } } }, + }, + '#withDisableMountSubPath':: d.fn(help='"*Deprecated: subPath usage will be removed in a future release.*"', args=[d.arg(name='disableMountSubPath', type=d.T.boolean)]), + withDisableMountSubPath(disableMountSubPath): { spec+: { storage+: { disableMountSubPath: disableMountSubPath } } }, + }, + '#tolerations':: d.obj(help="\"If specified, the pod's tolerations.\""), + tolerations: { + '#withEffect':: d.fn(help='"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."', args=[d.arg(name='effect', type=d.T.string)]), + withEffect(effect): { effect: effect }, + '#withKey':: d.fn(help='"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withTolerationSeconds':: d.fn(help='"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."', args=[d.arg(name='tolerationSeconds', type=d.T.integer)]), + withTolerationSeconds(tolerationSeconds): { tolerationSeconds: tolerationSeconds }, + '#withValue':: d.fn(help='"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#topologySpreadConstraints':: d.obj(help="\"If specified, the pod's topology spread constraints.\""), + topologySpreadConstraints: { + '#labelSelector':: d.obj(help='"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#withMatchLabelKeys':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='matchLabelKeys', type=d.T.array)]), + withMatchLabelKeys(matchLabelKeys): { matchLabelKeys: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] }, + '#withMatchLabelKeysMixin':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchLabelKeys', type=d.T.array)]), + withMatchLabelKeysMixin(matchLabelKeys): { matchLabelKeys+: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] }, + '#withMaxSkew':: d.fn(help="\"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed.\"", args=[d.arg(name='maxSkew', type=d.T.integer)]), + withMaxSkew(maxSkew): { maxSkew: maxSkew }, + '#withMinDomains':: d.fn(help="\"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \\\"global minimum\\\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \\n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \\\"global minimum\\\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \\n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='minDomains', type=d.T.integer)]), + withMinDomains(minDomains): { minDomains: minDomains }, + '#withNodeAffinityPolicy':: d.fn(help="\"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \\n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag.\"", args=[d.arg(name='nodeAffinityPolicy', type=d.T.string)]), + withNodeAffinityPolicy(nodeAffinityPolicy): { nodeAffinityPolicy: nodeAffinityPolicy }, + '#withNodeTaintsPolicy':: d.fn(help='"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \\n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."', args=[d.arg(name='nodeTaintsPolicy', type=d.T.string)]), + withNodeTaintsPolicy(nodeTaintsPolicy): { nodeTaintsPolicy: nodeTaintsPolicy }, + '#withTopologyKey':: d.fn(help="\"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each \u003ckey, value\u003e as a \\\"bucket\\\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \\\"kubernetes.io/hostname\\\", each Node is a domain of that topology. And, if TopologyKey is \\\"topology.kubernetes.io/zone\\\", each zone is a domain of that topology. It's a required field.\"", args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + '#withWhenUnsatisfiable':: d.fn(help="\"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \\\"Unsatisfiable\\\" for an incoming pod if and only if every possible node assignment for that pod would violate \\\"MaxSkew\\\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field.\"", args=[d.arg(name='whenUnsatisfiable', type=d.T.string)]), + withWhenUnsatisfiable(whenUnsatisfiable): { whenUnsatisfiable: whenUnsatisfiable }, + }, + '#volumeMounts':: d.obj(help='"VolumeMounts allows configuration of additional VolumeMounts on the output StatefulSet definition. VolumeMounts specified will be appended to other VolumeMounts in the alertmanager container, that are generated as a result of StorageSpec objects."'), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#volumes':: d.obj(help='"Volumes allows configuration of additional volumes on the output StatefulSet definition. Volumes specified will be appended to other volumes that are generated as a result of StorageSpec objects."'), + volumes: { + '#awsElasticBlockStore':: d.obj(help="\"awsElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore\""), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { awsElasticBlockStore+: { fsType: fsType } }, + '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty)."', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { awsElasticBlockStore+: { partition: partition } }, + '#withReadOnly':: d.fn(help='"readOnly value true will force the readOnly setting in VolumeMounts. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { awsElasticBlockStore+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID is unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { awsElasticBlockStore+: { volumeID: volumeID } }, + }, + '#azureDisk':: d.obj(help='"azureDisk represents an Azure Data Disk mount on the host and bind mount to the pod."'), + azureDisk: { + '#withCachingMode':: d.fn(help='"cachingMode is the Host Caching mode: None, Read Only, Read Write."', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { azureDisk+: { cachingMode: cachingMode } }, + '#withDiskName':: d.fn(help='"diskName is the Name of the data disk in the blob storage"', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { azureDisk+: { diskName: diskName } }, + '#withDiskURI':: d.fn(help='"diskURI is the URI of data disk in the blob storage"', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { azureDisk+: { diskURI: diskURI } }, + '#withFsType':: d.fn(help='"fsType is Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { azureDisk+: { fsType: fsType } }, + '#withKind':: d.fn(help='"kind expected values are Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { azureDisk+: { kind: kind } }, + '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureDisk+: { readOnly: readOnly } }, + }, + '#azureFile':: d.obj(help='"azureFile represents an Azure File Service mount on the host and bind mount to the pod."'), + azureFile: { + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureFile+: { readOnly: readOnly } }, + '#withSecretName':: d.fn(help='"secretName is the name of secret that contains Azure Storage Account Name and Key"', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { azureFile+: { secretName: secretName } }, + '#withShareName':: d.fn(help='"shareName is the azure share Name"', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { azureFile+: { shareName: shareName } }, + }, + '#cephfs':: d.obj(help="\"cephFS represents a Ceph FS mount on the host that shares a pod's lifetime\""), + cephfs: { + '#secretRef':: d.obj(help='"secretRef is Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cephfs+: { secretRef+: { name: name } } }, + }, + '#withMonitors':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPath':: d.fn(help='"path is Optional: Used as the mounted root, rather than the full Ceph tree, default is /"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { cephfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cephfs+: { readOnly: readOnly } }, + '#withSecretFile':: d.fn(help='"secretFile is Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { cephfs+: { secretFile: secretFile } }, + '#withUser':: d.fn(help='"user is optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { cephfs+: { user: user } }, + }, + '#cinder':: d.obj(help='"cinder represents a cinder volume attached and mounted on kubelets host machine. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"'), + cinder: { + '#secretRef':: d.obj(help='"secretRef is optional: points to a secret object containing parameters used to connect to OpenStack."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cinder+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { cinder+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cinder+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { cinder+: { volumeID: volumeID } }, + }, + '#configMap':: d.obj(help='"configMap represents a configMap that should populate this volume"'), + configMap: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode is optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { configMap+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#csi':: d.obj(help='"csi (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature)."'), + csi: { + '#nodePublishSecretRef':: d.obj(help='"nodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed."'), + nodePublishSecretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { nodePublishSecretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='"driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster."', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { csi+: { driver: driver } }, + '#withFsType':: d.fn(help='"fsType to mount. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { csi+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly specifies a read-only configuration for the volume. Defaults to false (read/write)."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { csi+: { readOnly: readOnly } }, + '#withVolumeAttributes':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { csi+: { volumeAttributes: volumeAttributes } }, + '#withVolumeAttributesMixin':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { csi+: { volumeAttributes+: volumeAttributes } }, + }, + '#downwardAPI':: d.obj(help='"downwardAPI represents downward API about the pod that should populate this volume"'), + downwardAPI: { + '#items':: d.obj(help='"Items is a list of downward API volume file"'), + items: { + '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { downwardAPI+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help='"Items is a list of downward API volume file"', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='"Items is a list of downward API volume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#emptyDir':: d.obj(help="\"emptyDir represents a temporary directory that shares a pod's lifetime. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\""), + emptyDir: { + '#withMedium':: d.fn(help="\"medium represents what type of storage medium should back this directory. The default is \\\"\\\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\"", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { emptyDir+: { medium: medium } }, + '#withSizeLimit':: d.fn(help='"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"', args=[d.arg(name='sizeLimit', type=d.T.any)]), + withSizeLimit(sizeLimit): { emptyDir+: { sizeLimit: sizeLimit } }, + }, + '#ephemeral':: d.obj(help="\"ephemeral represents a volume that is handled by a cluster storage driver. The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, and deleted when the pod is removed. \\n Use this if: a) the volume is only needed while the pod runs, b) features of normal volumes like restoring from snapshot or capacity tracking are needed, c) the storage driver is specified through a storage class, and d) the storage driver supports dynamic volume provisioning through a PersistentVolumeClaim (see EphemeralVolumeSource for more information on the connection between this volume type and PersistentVolumeClaim). \\n Use PersistentVolumeClaim or one of the vendor-specific APIs for volumes that persist for longer than the lifecycle of an individual pod. \\n Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to be used that way - see the documentation of the driver for more information. \\n A pod can use both types of ephemeral volumes and persistent volumes at the same time.\""), + ephemeral: { + '#volumeClaimTemplate':: d.obj(help='"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \\n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \\n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \\n Required, must not be nil."'), + volumeClaimTemplate: { + '#spec':: d.obj(help='"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } }, + }, + '#withMetadata':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadata(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata: metadata } } }, + '#withMetadataMixin':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadataMixin(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata+: metadata } } }, + }, + }, + '#fc':: d.obj(help="\"fc represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod.\""), + fc: { + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fc+: { fsType: fsType } }, + '#withLun':: d.fn(help='"lun is Optional: FC target lun number"', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { fc+: { lun: lun } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { fc+: { readOnly: readOnly } }, + '#withTargetWWNs':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withTargetWWNsMixin':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withWwids':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } }, + '#withWwidsMixin':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } }, + }, + '#flexVolume':: d.obj(help='"flexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin."'), + flexVolume: { + '#secretRef':: d.obj(help='"secretRef is Optional: secretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { flexVolume+: { secretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='"driver is the name of the driver to use for this volume."', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { flexVolume+: { driver: driver } }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". The default filesystem depends on FlexVolume script."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { flexVolume+: { fsType: fsType } }, + '#withOptions':: d.fn(help='"options is Optional: this field holds extra command options if any."', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { flexVolume+: { options: options } }, + '#withOptionsMixin':: d.fn(help='"options is Optional: this field holds extra command options if any."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { flexVolume+: { options+: options } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { flexVolume+: { readOnly: readOnly } }, + }, + '#flocker':: d.obj(help="\"flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running\""), + flocker: { + '#withDatasetName':: d.fn(help='"datasetName is Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated"', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { flocker+: { datasetName: datasetName } }, + '#withDatasetUUID':: d.fn(help='"datasetUUID is the UUID of the dataset. This is unique identifier of a Flocker dataset"', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { flocker+: { datasetUUID: datasetUUID } }, + }, + '#gcePersistentDisk':: d.obj(help="\"gcePersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk\""), + gcePersistentDisk: { + '#withFsType':: d.fn(help='"fsType is filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { gcePersistentDisk+: { fsType: fsType } }, + '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { gcePersistentDisk+: { partition: partition } }, + '#withPdName':: d.fn(help='"pdName is unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { gcePersistentDisk+: { pdName: pdName } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { gcePersistentDisk+: { readOnly: readOnly } }, + }, + '#gitRepo':: d.obj(help="\"gitRepo represents a git repository at a particular revision. DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container.\""), + gitRepo: { + '#withDirectory':: d.fn(help="\"directory is the target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name.\"", args=[d.arg(name='directory', type=d.T.string)]), + withDirectory(directory): { gitRepo+: { directory: directory } }, + '#withRepository':: d.fn(help='"repository is the URL"', args=[d.arg(name='repository', type=d.T.string)]), + withRepository(repository): { gitRepo+: { repository: repository } }, + '#withRevision':: d.fn(help='"revision is the commit hash for the specified revision."', args=[d.arg(name='revision', type=d.T.string)]), + withRevision(revision): { gitRepo+: { revision: revision } }, + }, + '#glusterfs':: d.obj(help="\"glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/glusterfs/README.md\""), + glusterfs: { + '#withEndpoints':: d.fn(help='"endpoints is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { glusterfs+: { endpoints: endpoints } }, + '#withPath':: d.fn(help='"path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { glusterfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { glusterfs+: { readOnly: readOnly } }, + }, + '#hostPath':: d.obj(help='"hostPath represents a pre-existing file or directory on the host machine that is directly exposed to the container. This is generally used for system agents or other privileged things that are allowed to see the host machine. Most containers will NOT need this. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath --- TODO(jonesdl) We need to restrict who can use host directory mounts and who can/can not mount host directories as read/write."'), + hostPath: { + '#withPath':: d.fn(help='"path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { hostPath+: { path: path } }, + '#withType':: d.fn(help='"type for HostPath Volume Defaults to \\"\\" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { hostPath+: { type: type } }, + }, + '#iscsi':: d.obj(help="\"iscsi represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://examples.k8s.io/volumes/iscsi/README.md\""), + iscsi: { + '#secretRef':: d.obj(help='"secretRef is the CHAP Secret for iSCSI target and initiator authentication"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { iscsi+: { secretRef+: { name: name } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='"chapAuthDiscovery defines whether support iSCSI Discovery CHAP authentication"', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } }, + '#withChapAuthSession':: d.fn(help='"chapAuthSession defines whether support iSCSI Session CHAP authentication"', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { iscsi+: { chapAuthSession: chapAuthSession } }, + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { iscsi+: { fsType: fsType } }, + '#withInitiatorName':: d.fn(help='"initiatorName is the custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection."', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { iscsi+: { initiatorName: initiatorName } }, + '#withIqn':: d.fn(help='"iqn is the target iSCSI Qualified Name."', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iscsi+: { iqn: iqn } }, + '#withIscsiInterface':: d.fn(help="\"iscsiInterface is the interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).\"", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsi+: { iscsiInterface: iscsiInterface } }, + '#withLun':: d.fn(help='"lun represents iSCSI Target Lun number."', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { iscsi+: { lun: lun } }, + '#withPortals':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } }, + '#withPortalsMixin':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { iscsi+: { readOnly: readOnly } }, + '#withTargetPortal':: d.fn(help='"targetPortal is iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { iscsi+: { targetPortal: targetPortal } }, + }, + '#nfs':: d.obj(help="\"nfs represents an NFS mount on the host that shares a pod's lifetime More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs\""), + nfs: { + '#withPath':: d.fn(help='"path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { nfs+: { readOnly: readOnly } }, + '#withServer':: d.fn(help='"server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { nfs+: { server: server } }, + }, + '#persistentVolumeClaim':: d.obj(help='"persistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"'), + persistentVolumeClaim: { + '#withClaimName':: d.fn(help='"claimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"', args=[d.arg(name='claimName', type=d.T.string)]), + withClaimName(claimName): { persistentVolumeClaim+: { claimName: claimName } }, + '#withReadOnly':: d.fn(help='"readOnly Will force the ReadOnly setting in VolumeMounts. Default false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { persistentVolumeClaim+: { readOnly: readOnly } }, + }, + '#photonPersistentDisk':: d.obj(help='"photonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine"'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { photonPersistentDisk+: { fsType: fsType } }, + '#withPdID':: d.fn(help='"pdID is the ID that identifies Photon Controller persistent disk"', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { photonPersistentDisk+: { pdID: pdID } }, + }, + '#portworxVolume':: d.obj(help='"portworxVolume represents a portworx volume attached and mounted on kubelets host machine"'), + portworxVolume: { + '#withFsType':: d.fn(help='"fSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { portworxVolume+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { portworxVolume+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID uniquely identifies a Portworx volume"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { portworxVolume+: { volumeID: volumeID } }, + }, + '#projected':: d.obj(help='"projected items for all in one resources secrets, configmaps, and downward API"'), + projected: { + '#sources':: d.obj(help='"sources is the list of volume projections"'), + sources: { + '#configMap':: d.obj(help='"configMap information about the configMap data to project"'), + configMap: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#downwardAPI':: d.obj(help='"downwardAPI information about the downwardAPI data to project"'), + downwardAPI: { + '#items':: d.obj(help='"Items is a list of DownwardAPIVolume file"'), + items: { + '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help='"Items is a list of DownwardAPIVolume file"', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='"Items is a list of DownwardAPIVolume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#secret':: d.obj(help='"secret information about the secret data to project"'), + secret: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secret+: { name: name } }, + '#withOptional':: d.fn(help='"optional field specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + }, + '#serviceAccountToken':: d.obj(help='"serviceAccountToken is information about the serviceAccountToken data to project"'), + serviceAccountToken: { + '#withAudience':: d.fn(help='"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."', args=[d.arg(name='audience', type=d.T.string)]), + withAudience(audience): { serviceAccountToken+: { audience: audience } }, + '#withExpirationSeconds':: d.fn(help='"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."', args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { serviceAccountToken+: { expirationSeconds: expirationSeconds } }, + '#withPath':: d.fn(help='"path is the path relative to the mount point of the file to project the token into."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { serviceAccountToken+: { path: path } }, + }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { projected+: { defaultMode: defaultMode } }, + '#withSources':: d.fn(help='"sources is the list of volume projections"', args=[d.arg(name='sources', type=d.T.array)]), + withSources(sources): { projected+: { sources: if std.isArray(v=sources) then sources else [sources] } }, + '#withSourcesMixin':: d.fn(help='"sources is the list of volume projections"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]), + withSourcesMixin(sources): { projected+: { sources+: if std.isArray(v=sources) then sources else [sources] } }, + }, + '#quobyte':: d.obj(help="\"quobyte represents a Quobyte mount on the host that shares a pod's lifetime\""), + quobyte: { + '#withGroup':: d.fn(help='"group to map volume access to Default is no group"', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { quobyte+: { group: group } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { quobyte+: { readOnly: readOnly } }, + '#withRegistry':: d.fn(help='"registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes"', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { quobyte+: { registry: registry } }, + '#withTenant':: d.fn(help='"tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin"', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { quobyte+: { tenant: tenant } }, + '#withUser':: d.fn(help='"user to map volume access to Defaults to serivceaccount user"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { quobyte+: { user: user } }, + '#withVolume':: d.fn(help='"volume is a string that references an already created Quobyte volume by name."', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { quobyte+: { volume: volume } }, + }, + '#rbd':: d.obj(help="\"rbd represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/rbd/README.md\""), + rbd: { + '#secretRef':: d.obj(help='"secretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { rbd+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { rbd+: { fsType: fsType } }, + '#withImage':: d.fn(help='"image is the rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { rbd+: { image: image } }, + '#withKeyring':: d.fn(help='"keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { rbd+: { keyring: keyring } }, + '#withMonitors':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPool':: d.fn(help='"pool is the rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { rbd+: { pool: pool } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { rbd+: { readOnly: readOnly } }, + '#withUser':: d.fn(help='"user is the rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { rbd+: { user: user } }, + }, + '#scaleIO':: d.obj(help='"scaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes."'), + scaleIO: { + '#secretRef':: d.obj(help='"secretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleIO+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Default is \\"xfs\\"."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { scaleIO+: { fsType: fsType } }, + '#withGateway':: d.fn(help='"gateway is the host address of the ScaleIO API Gateway."', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { scaleIO+: { gateway: gateway } }, + '#withProtectionDomain':: d.fn(help='"protectionDomain is the name of the ScaleIO Protection Domain for the configured storage."', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { scaleIO+: { protectionDomain: protectionDomain } }, + '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { scaleIO+: { readOnly: readOnly } }, + '#withSslEnabled':: d.fn(help='"sslEnabled Flag enable/disable SSL communication with Gateway, default false"', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { scaleIO+: { sslEnabled: sslEnabled } }, + '#withStorageMode':: d.fn(help='"storageMode indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned."', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { scaleIO+: { storageMode: storageMode } }, + '#withStoragePool':: d.fn(help='"storagePool is the ScaleIO Storage Pool associated with the protection domain."', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { scaleIO+: { storagePool: storagePool } }, + '#withSystem':: d.fn(help='"system is the name of the storage system as configured in ScaleIO."', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { scaleIO+: { system: system } }, + '#withVolumeName':: d.fn(help='"volumeName is the name of a volume already created in the ScaleIO system that is associated with this volume source."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { scaleIO+: { volumeName: volumeName } }, + }, + '#secret':: d.obj(help='"secret represents a secret that should populate this volume. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret"'), + secret: { + '#items':: d.obj(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode is Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { secret+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withOptional':: d.fn(help='"optional field specify whether the Secret or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + '#withSecretName':: d.fn(help="\"secretName is the name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret\"", args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secret+: { secretName: secretName } }, + }, + '#storageos':: d.obj(help='"storageOS represents a StorageOS volume attached and mounted on Kubernetes nodes."'), + storageos: { + '#secretRef':: d.obj(help='"secretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { storageos+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { storageos+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { storageos+: { readOnly: readOnly } }, + '#withVolumeName':: d.fn(help='"volumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { storageos+: { volumeName: volumeName } }, + '#withVolumeNamespace':: d.fn(help="\"volumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to \\\"default\\\" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.\"", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { storageos+: { volumeNamespace: volumeNamespace } }, + }, + '#vsphereVolume':: d.obj(help='"vsphereVolume represents a vSphere volume attached and mounted on kubelets host machine"'), + vsphereVolume: { + '#withFsType':: d.fn(help='"fsType is filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { vsphereVolume+: { fsType: fsType } }, + '#withStoragePolicyID':: d.fn(help='"storagePolicyID is the storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName."', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { vsphereVolume+: { storagePolicyID: storagePolicyID } }, + '#withStoragePolicyName':: d.fn(help='"storagePolicyName is the storage Policy Based Management (SPBM) profile name."', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { vsphereVolume+: { storagePolicyName: storagePolicyName } }, + '#withVolumePath':: d.fn(help='"volumePath is the path that identifies vSphere volume vmdk"', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { vsphereVolume+: { volumePath: volumePath } }, + }, + '#withName':: d.fn(help='"name of the volume. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#web':: d.obj(help='"Defines the web command line flags when starting Alertmanager."'), + web: { + '#httpConfig':: d.obj(help='"Defines HTTP parameters for web server."'), + httpConfig: { + '#headers':: d.obj(help='"List of headers that can be added to HTTP responses."'), + headers: { + '#withContentSecurityPolicy':: d.fn(help='"Set the Content-Security-Policy header to HTTP responses. Unset if blank."', args=[d.arg(name='contentSecurityPolicy', type=d.T.string)]), + withContentSecurityPolicy(contentSecurityPolicy): { spec+: { web+: { httpConfig+: { headers+: { contentSecurityPolicy: contentSecurityPolicy } } } } }, + '#withStrictTransportSecurity':: d.fn(help='"Set the Strict-Transport-Security header to HTTP responses. Unset if blank. Please make sure that you use this with care as this header might force browsers to load Prometheus and the other applications hosted on the same domain and subdomains over HTTPS. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Strict-Transport-Security"', args=[d.arg(name='strictTransportSecurity', type=d.T.string)]), + withStrictTransportSecurity(strictTransportSecurity): { spec+: { web+: { httpConfig+: { headers+: { strictTransportSecurity: strictTransportSecurity } } } } }, + '#withXContentTypeOptions':: d.fn(help='"Set the X-Content-Type-Options header to HTTP responses. Unset if blank. Accepted value is nosniff. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Content-Type-Options"', args=[d.arg(name='xContentTypeOptions', type=d.T.string)]), + withXContentTypeOptions(xContentTypeOptions): { spec+: { web+: { httpConfig+: { headers+: { xContentTypeOptions: xContentTypeOptions } } } } }, + '#withXFrameOptions':: d.fn(help='"Set the X-Frame-Options header to HTTP responses. Unset if blank. Accepted values are deny and sameorigin. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options"', args=[d.arg(name='xFrameOptions', type=d.T.string)]), + withXFrameOptions(xFrameOptions): { spec+: { web+: { httpConfig+: { headers+: { xFrameOptions: xFrameOptions } } } } }, + '#withXXSSProtection':: d.fn(help='"Set the X-XSS-Protection header to all responses. Unset if blank. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-XSS-Protection"', args=[d.arg(name='xXSSProtection', type=d.T.string)]), + withXXSSProtection(xXSSProtection): { spec+: { web+: { httpConfig+: { headers+: { xXSSProtection: xXSSProtection } } } } }, + }, + '#withHttp2':: d.fn(help='"Enable HTTP/2 support. Note that HTTP/2 is only supported with TLS. When TLSConfig is not configured, HTTP/2 will be disabled. Whenever the value of the field changes, a rolling update will be triggered."', args=[d.arg(name='http2', type=d.T.boolean)]), + withHttp2(http2): { spec+: { web+: { httpConfig+: { http2: http2 } } } }, + }, + '#tlsConfig':: d.obj(help='"Defines the TLS parameters for HTTPS."'), + tlsConfig: { + '#cert':: d.obj(help='"Contains the TLS certificate for the server."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } } }, + }, + }, + '#client_ca':: d.obj(help='"Contains the CA certificate for client certificate authentication to the server."'), + client_ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { optional: optional } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the TLS key for the server."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { keySecret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { keySecret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { keySecret+: { optional: optional } } } } }, + }, + '#withCipherSuites':: d.fn(help='"List of supported cipher suites for TLS versions up to TLS 1.2. If empty, Go default cipher suites are used. Available cipher suites are documented in the go documentation: https://golang.org/pkg/crypto/tls/#pkg-constants"', args=[d.arg(name='cipherSuites', type=d.T.array)]), + withCipherSuites(cipherSuites): { spec+: { web+: { tlsConfig+: { cipherSuites: if std.isArray(v=cipherSuites) then cipherSuites else [cipherSuites] } } } }, + '#withCipherSuitesMixin':: d.fn(help='"List of supported cipher suites for TLS versions up to TLS 1.2. If empty, Go default cipher suites are used. Available cipher suites are documented in the go documentation: https://golang.org/pkg/crypto/tls/#pkg-constants"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='cipherSuites', type=d.T.array)]), + withCipherSuitesMixin(cipherSuites): { spec+: { web+: { tlsConfig+: { cipherSuites+: if std.isArray(v=cipherSuites) then cipherSuites else [cipherSuites] } } } }, + '#withClientAuthType':: d.fn(help='"Server policy for client authentication. Maps to ClientAuth Policies. For more detail on clientAuth options: https://golang.org/pkg/crypto/tls/#ClientAuthType"', args=[d.arg(name='clientAuthType', type=d.T.string)]), + withClientAuthType(clientAuthType): { spec+: { web+: { tlsConfig+: { clientAuthType: clientAuthType } } } }, + '#withCurvePreferences':: d.fn(help='"Elliptic curves that will be used in an ECDHE handshake, in preference order. Available curves are documented in the go documentation: https://golang.org/pkg/crypto/tls/#CurveID"', args=[d.arg(name='curvePreferences', type=d.T.array)]), + withCurvePreferences(curvePreferences): { spec+: { web+: { tlsConfig+: { curvePreferences: if std.isArray(v=curvePreferences) then curvePreferences else [curvePreferences] } } } }, + '#withCurvePreferencesMixin':: d.fn(help='"Elliptic curves that will be used in an ECDHE handshake, in preference order. Available curves are documented in the go documentation: https://golang.org/pkg/crypto/tls/#CurveID"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='curvePreferences', type=d.T.array)]), + withCurvePreferencesMixin(curvePreferences): { spec+: { web+: { tlsConfig+: { curvePreferences+: if std.isArray(v=curvePreferences) then curvePreferences else [curvePreferences] } } } }, + '#withMaxVersion':: d.fn(help='"Maximum TLS version that is acceptable. Defaults to TLS13."', args=[d.arg(name='maxVersion', type=d.T.string)]), + withMaxVersion(maxVersion): { spec+: { web+: { tlsConfig+: { maxVersion: maxVersion } } } }, + '#withMinVersion':: d.fn(help='"Minimum TLS version that is acceptable. Defaults to TLS12."', args=[d.arg(name='minVersion', type=d.T.string)]), + withMinVersion(minVersion): { spec+: { web+: { tlsConfig+: { minVersion: minVersion } } } }, + '#withPreferServerCipherSuites':: d.fn(help="\"Controls whether the server selects the client's most preferred cipher suite, or the server's most preferred cipher suite. If true then the server's preference, as expressed in the order of elements in cipherSuites, is used.\"", args=[d.arg(name='preferServerCipherSuites', type=d.T.boolean)]), + withPreferServerCipherSuites(preferServerCipherSuites): { spec+: { web+: { tlsConfig+: { preferServerCipherSuites: preferServerCipherSuites } } } }, + }, + '#withGetConcurrency':: d.fn(help="\"Maximum number of GET requests processed concurrently. This corresponds to the Alertmanager's `--web.get-concurrency` flag.\"", args=[d.arg(name='getConcurrency', type=d.T.integer)]), + withGetConcurrency(getConcurrency): { spec+: { web+: { getConcurrency: getConcurrency } } }, + '#withTimeout':: d.fn(help="\"Timeout for HTTP requests. This corresponds to the Alertmanager's `--web.timeout` flag.\"", args=[d.arg(name='timeout', type=d.T.integer)]), + withTimeout(timeout): { spec+: { web+: { timeout: timeout } } }, + }, + '#withAdditionalPeers':: d.fn(help='"AdditionalPeers allows injecting a set of additional Alertmanagers to peer with to form a highly available cluster."', args=[d.arg(name='additionalPeers', type=d.T.array)]), + withAdditionalPeers(additionalPeers): { spec+: { additionalPeers: if std.isArray(v=additionalPeers) then additionalPeers else [additionalPeers] } }, + '#withAdditionalPeersMixin':: d.fn(help='"AdditionalPeers allows injecting a set of additional Alertmanagers to peer with to form a highly available cluster."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='additionalPeers', type=d.T.array)]), + withAdditionalPeersMixin(additionalPeers): { spec+: { additionalPeers+: if std.isArray(v=additionalPeers) then additionalPeers else [additionalPeers] } }, + '#withAutomountServiceAccountToken':: d.fn(help='"AutomountServiceAccountToken indicates whether a service account token should be automatically mounted in the pod. If the service account has `automountServiceAccountToken: true`, set the field to `false` to opt out of automounting API credentials."', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]), + withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { automountServiceAccountToken: automountServiceAccountToken } }, + '#withBaseImage':: d.fn(help="\"Base image that is used to deploy pods, without tag. Deprecated: use 'image' instead\"", args=[d.arg(name='baseImage', type=d.T.string)]), + withBaseImage(baseImage): { spec+: { baseImage: baseImage } }, + '#withClusterAdvertiseAddress':: d.fn(help='"ClusterAdvertiseAddress is the explicit address to advertise in cluster. Needs to be provided for non RFC1918 [1] (public) addresses. [1] RFC1918: https://tools.ietf.org/html/rfc1918"', args=[d.arg(name='clusterAdvertiseAddress', type=d.T.string)]), + withClusterAdvertiseAddress(clusterAdvertiseAddress): { spec+: { clusterAdvertiseAddress: clusterAdvertiseAddress } }, + '#withClusterGossipInterval':: d.fn(help='"Interval between gossip attempts."', args=[d.arg(name='clusterGossipInterval', type=d.T.string)]), + withClusterGossipInterval(clusterGossipInterval): { spec+: { clusterGossipInterval: clusterGossipInterval } }, + '#withClusterPeerTimeout':: d.fn(help='"Timeout for cluster peering."', args=[d.arg(name='clusterPeerTimeout', type=d.T.string)]), + withClusterPeerTimeout(clusterPeerTimeout): { spec+: { clusterPeerTimeout: clusterPeerTimeout } }, + '#withClusterPushpullInterval':: d.fn(help='"Interval between pushpull attempts."', args=[d.arg(name='clusterPushpullInterval', type=d.T.string)]), + withClusterPushpullInterval(clusterPushpullInterval): { spec+: { clusterPushpullInterval: clusterPushpullInterval } }, + '#withConfigMaps':: d.fn(help="\"ConfigMaps is a list of ConfigMaps in the same namespace as the Alertmanager object, which shall be mounted into the Alertmanager Pods. Each ConfigMap is added to the StatefulSet definition as a volume named `configmap-\u003cconfigmap-name\u003e`. The ConfigMaps are mounted into `/etc/alertmanager/configmaps/\u003cconfigmap-name\u003e` in the 'alertmanager' container.\"", args=[d.arg(name='configMaps', type=d.T.array)]), + withConfigMaps(configMaps): { spec+: { configMaps: if std.isArray(v=configMaps) then configMaps else [configMaps] } }, + '#withConfigMapsMixin':: d.fn(help="\"ConfigMaps is a list of ConfigMaps in the same namespace as the Alertmanager object, which shall be mounted into the Alertmanager Pods. Each ConfigMap is added to the StatefulSet definition as a volume named `configmap-\u003cconfigmap-name\u003e`. The ConfigMaps are mounted into `/etc/alertmanager/configmaps/\u003cconfigmap-name\u003e` in the 'alertmanager' container.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='configMaps', type=d.T.array)]), + withConfigMapsMixin(configMaps): { spec+: { configMaps+: if std.isArray(v=configMaps) then configMaps else [configMaps] } }, + '#withConfigSecret':: d.fn(help='"ConfigSecret is the name of a Kubernetes Secret in the same namespace as the Alertmanager object, which contains the configuration for this Alertmanager instance. If empty, it defaults to `alertmanager-`. \\n The Alertmanager configuration should be available under the `alertmanager.yaml` key. Additional keys from the original secret are copied to the generated secret and mounted into the `/etc/alertmanager/config` directory in the `alertmanager` container. \\n If either the secret or the `alertmanager.yaml` key is missing, the operator provisions a minimal Alertmanager configuration with one empty receiver (effectively dropping alert notifications)."', args=[d.arg(name='configSecret', type=d.T.string)]), + withConfigSecret(configSecret): { spec+: { configSecret: configSecret } }, + '#withContainers':: d.fn(help='"Containers allows injecting additional containers. This is meant to allow adding an authentication proxy to an Alertmanager pod. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. The current container names are: `alertmanager` and `config-reloader`. Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } }, + '#withContainersMixin':: d.fn(help='"Containers allows injecting additional containers. This is meant to allow adding an authentication proxy to an Alertmanager pod. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. The current container names are: `alertmanager` and `config-reloader`. Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } }, + '#withExternalUrl':: d.fn(help='"The external URL the Alertmanager instances will be available under. This is necessary to generate correct URLs. This is necessary if Alertmanager is not served from root of a DNS name."', args=[d.arg(name='externalUrl', type=d.T.string)]), + withExternalUrl(externalUrl): { spec+: { externalUrl: externalUrl } }, + '#withForceEnableClusterMode':: d.fn(help='"ForceEnableClusterMode ensures Alertmanager does not deactivate the cluster mode when running with a single replica. Use case is e.g. spanning an Alertmanager cluster across Kubernetes clusters with a single replica in each."', args=[d.arg(name='forceEnableClusterMode', type=d.T.boolean)]), + withForceEnableClusterMode(forceEnableClusterMode): { spec+: { forceEnableClusterMode: forceEnableClusterMode } }, + '#withHostAliases':: d.fn(help="\"Pods' hostAliases configuration\"", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostAliasesMixin':: d.fn(help="\"Pods' hostAliases configuration\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withImage':: d.fn(help='"Image if specified has precedence over baseImage, tag and sha combinations. Specifying the version is still necessary to ensure the Prometheus Operator knows what version of Alertmanager is being configured."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { image: image } }, + '#withImagePullPolicy':: d.fn(help="\"Image pull policy for the 'alertmanager', 'init-config-reloader' and 'config-reloader' containers. See https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy for more details.\"", args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { spec+: { imagePullPolicy: imagePullPolicy } }, + '#withImagePullSecrets':: d.fn(help='"An optional list of references to secrets in the same namespace to use for pulling prometheus and alertmanager images from registries see http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withImagePullSecretsMixin':: d.fn(help='"An optional list of references to secrets in the same namespace to use for pulling prometheus and alertmanager images from registries see http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withInitContainers':: d.fn(help='"InitContainers allows adding initContainers to the pod definition. Those can be used to e.g. fetch secrets for injection into the Alertmanager configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. The current init container name is: `init-config-reloader`. Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withInitContainersMixin':: d.fn(help='"InitContainers allows adding initContainers to the pod definition. Those can be used to e.g. fetch secrets for injection into the Alertmanager configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. The current init container name is: `init-config-reloader`. Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withListenLocal':: d.fn(help='"ListenLocal makes the Alertmanager server listen on loopback, so that it does not bind against the Pod IP. Note this is only for the Alertmanager UI, not the gossip communication."', args=[d.arg(name='listenLocal', type=d.T.boolean)]), + withListenLocal(listenLocal): { spec+: { listenLocal: listenLocal } }, + '#withLogFormat':: d.fn(help='"Log format for Alertmanager to be configured with."', args=[d.arg(name='logFormat', type=d.T.string)]), + withLogFormat(logFormat): { spec+: { logFormat: logFormat } }, + '#withLogLevel':: d.fn(help='"Log level for Alertmanager to be configured with."', args=[d.arg(name='logLevel', type=d.T.string)]), + withLogLevel(logLevel): { spec+: { logLevel: logLevel } }, + '#withMinReadySeconds':: d.fn(help='"Minimum number of seconds for which a newly created pod should be ready without any of its container crashing for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) This is an alpha field from kubernetes 1.22 until 1.24 which requires enabling the StatefulSetMinReadySeconds feature gate."', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withNodeSelector':: d.fn(help='"Define which Nodes the Pods are scheduled on."', args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help='"Define which Nodes the Pods are scheduled on."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { nodeSelector+: nodeSelector } }, + '#withPaused':: d.fn(help='"If set to true all actions on the underlying managed objects are not goint to be performed, except for delete actions."', args=[d.arg(name='paused', type=d.T.boolean)]), + withPaused(paused): { spec+: { paused: paused } }, + '#withPortName':: d.fn(help='"Port name used for the pods and governing service. Defaults to `web`."', args=[d.arg(name='portName', type=d.T.string)]), + withPortName(portName): { spec+: { portName: portName } }, + '#withPriorityClassName':: d.fn(help='"Priority class assigned to the Pods"', args=[d.arg(name='priorityClassName', type=d.T.string)]), + withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } }, + '#withReplicas':: d.fn(help='"Size is the expected size of the alertmanager cluster. The controller will eventually make the size of the running cluster equal to the expected size."', args=[d.arg(name='replicas', type=d.T.integer)]), + withReplicas(replicas): { spec+: { replicas: replicas } }, + '#withRetention':: d.fn(help="\"Time duration Alertmanager shall retain data for. Default is '120h', and must match the regular expression `[0-9]+(ms|s|m|h)` (milliseconds seconds minutes hours).\"", args=[d.arg(name='retention', type=d.T.string)]), + withRetention(retention): { spec+: { retention: retention } }, + '#withRoutePrefix':: d.fn(help='"The route prefix Alertmanager registers HTTP handlers for. This is useful, if using ExternalURL and a proxy is rewriting HTTP routes of a request, and the actual ExternalURL is still true, but the server serves requests under a different route prefix. For example for use with `kubectl proxy`."', args=[d.arg(name='routePrefix', type=d.T.string)]), + withRoutePrefix(routePrefix): { spec+: { routePrefix: routePrefix } }, + '#withSecrets':: d.fn(help="\"Secrets is a list of Secrets in the same namespace as the Alertmanager object, which shall be mounted into the Alertmanager Pods. Each Secret is added to the StatefulSet definition as a volume named `secret-\u003csecret-name\u003e`. The Secrets are mounted into `/etc/alertmanager/secrets/\u003csecret-name\u003e` in the 'alertmanager' container.\"", args=[d.arg(name='secrets', type=d.T.array)]), + withSecrets(secrets): { spec+: { secrets: if std.isArray(v=secrets) then secrets else [secrets] } }, + '#withSecretsMixin':: d.fn(help="\"Secrets is a list of Secrets in the same namespace as the Alertmanager object, which shall be mounted into the Alertmanager Pods. Each Secret is added to the StatefulSet definition as a volume named `secret-\u003csecret-name\u003e`. The Secrets are mounted into `/etc/alertmanager/secrets/\u003csecret-name\u003e` in the 'alertmanager' container.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='secrets', type=d.T.array)]), + withSecretsMixin(secrets): { spec+: { secrets+: if std.isArray(v=secrets) then secrets else [secrets] } }, + '#withServiceAccountName':: d.fn(help='"ServiceAccountName is the name of the ServiceAccount to use to run the Prometheus Pods."', args=[d.arg(name='serviceAccountName', type=d.T.string)]), + withServiceAccountName(serviceAccountName): { spec+: { serviceAccountName: serviceAccountName } }, + '#withSha':: d.fn(help="\"SHA of Alertmanager container image to be deployed. Defaults to the value of `version`. Similar to a tag, but the SHA explicitly deploys an immutable container image. Version and Tag are ignored if SHA is set. Deprecated: use 'image' instead. The image digest can be specified as part of the image URL.\"", args=[d.arg(name='sha', type=d.T.string)]), + withSha(sha): { spec+: { sha: sha } }, + '#withTag':: d.fn(help="\"Tag of Alertmanager container image to be deployed. Defaults to the value of `version`. Version is ignored if Tag is set. Deprecated: use 'image' instead. The image tag can be specified as part of the image URL.\"", args=[d.arg(name='tag', type=d.T.string)]), + withTag(tag): { spec+: { tag: tag } }, + '#withTolerations':: d.fn(help="\"If specified, the pod's tolerations.\"", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerations(tolerations): { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTolerationsMixin':: d.fn(help="\"If specified, the pod's tolerations.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]), + withTolerationsMixin(tolerations): { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } }, + '#withTopologySpreadConstraints':: d.fn(help="\"If specified, the pod's topology spread constraints.\"", args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withTopologySpreadConstraintsMixin':: d.fn(help="\"If specified, the pod's topology spread constraints.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]), + withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } }, + '#withVersion':: d.fn(help='"Version the cluster should be on."', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { version: version } }, + '#withVolumeMounts':: d.fn(help='"VolumeMounts allows configuration of additional VolumeMounts on the output StatefulSet definition. VolumeMounts specified will be appended to other VolumeMounts in the alertmanager container, that are generated as a result of StorageSpec objects."', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { spec+: { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } }, + '#withVolumeMountsMixin':: d.fn(help='"VolumeMounts allows configuration of additional VolumeMounts on the output StatefulSet definition. VolumeMounts specified will be appended to other VolumeMounts in the alertmanager container, that are generated as a result of StorageSpec objects."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { spec+: { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } }, + '#withVolumes':: d.fn(help='"Volumes allows configuration of additional volumes on the output StatefulSet definition. Volumes specified will be appended to other volumes that are generated as a result of StorageSpec objects."', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumes(volumes): { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } }, + '#withVolumesMixin':: d.fn(help='"Volumes allows configuration of additional volumes on the output StatefulSet definition. Volumes specified will be appended to other volumes that are generated as a result of StorageSpec objects."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]), + withVolumesMixin(volumes): { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/0.13/_gen/monitoring/v1/main.libsonnet b/0.13/_gen/monitoring/v1/main.libsonnet new file mode 100644 index 0000000..35ee044 --- /dev/null +++ b/0.13/_gen/monitoring/v1/main.libsonnet @@ -0,0 +1,11 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='v1', url='', help=''), + alertmanager: (import 'alertmanager.libsonnet'), + podMonitor: (import 'podMonitor.libsonnet'), + probe: (import 'probe.libsonnet'), + prometheus: (import 'prometheus.libsonnet'), + prometheusRule: (import 'prometheusRule.libsonnet'), + serviceMonitor: (import 'serviceMonitor.libsonnet'), + thanosRuler: (import 'thanosRuler.libsonnet'), +} diff --git a/0.13/_gen/monitoring/v1/podMonitor.libsonnet b/0.13/_gen/monitoring/v1/podMonitor.libsonnet new file mode 100644 index 0000000..e6cadec --- /dev/null +++ b/0.13/_gen/monitoring/v1/podMonitor.libsonnet @@ -0,0 +1,333 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='podMonitor', url='', help='"PodMonitor defines monitoring for a set of pods."'), + '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of PodMonitor', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'monitoring.coreos.com/v1', + kind: 'PodMonitor', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='"Specification of desired Pod selection for target discovery by Prometheus."'), + spec: { + '#attachMetadata':: d.obj(help='"Attaches node metadata to discovered targets. Requires Prometheus v2.35.0 and above."'), + attachMetadata: { + '#withNode':: d.fn(help='"When set to true, Prometheus must have permissions to get Nodes."', args=[d.arg(name='node', type=d.T.boolean)]), + withNode(node): { spec+: { attachMetadata+: { node: node } } }, + }, + '#namespaceSelector':: d.obj(help='"Selector to select which namespaces the Endpoints objects are discovered from."'), + namespaceSelector: { + '#withAny':: d.fn(help='"Boolean describing whether all namespaces are selected in contrast to a list restricting them."', args=[d.arg(name='any', type=d.T.boolean)]), + withAny(any): { spec+: { namespaceSelector+: { any: any } } }, + '#withMatchNames':: d.fn(help='"List of namespace names to select from."', args=[d.arg(name='matchNames', type=d.T.array)]), + withMatchNames(matchNames): { spec+: { namespaceSelector+: { matchNames: if std.isArray(v=matchNames) then matchNames else [matchNames] } } }, + '#withMatchNamesMixin':: d.fn(help='"List of namespace names to select from."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchNames', type=d.T.array)]), + withMatchNamesMixin(matchNames): { spec+: { namespaceSelector+: { matchNames+: if std.isArray(v=matchNames) then matchNames else [matchNames] } } }, + }, + '#podMetricsEndpoints':: d.obj(help='"A list of endpoints allowed as part of this PodMonitor."'), + podMetricsEndpoints: { + '#authorization':: d.obj(help='"Authorization section for this endpoint"'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { authorization+: { credentials+: { optional: optional } } }, + }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { authorization+: { type: type } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth allow an endpoint to authenticate over basic authentication. More info: https://prometheus.io/docs/operating/configuration/#endpoint"'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { password+: { optional: optional } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { username+: { optional: optional } } }, + }, + }, + '#bearerTokenSecret':: d.obj(help='"Secret to mount to read bearer token for scraping targets. The secret needs to be in the same namespace as the pod monitor and accessible by the Prometheus Operator."'), + bearerTokenSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { bearerTokenSecret+: { key: key } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { bearerTokenSecret+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { bearerTokenSecret+: { optional: optional } }, + }, + '#metricRelabelings':: d.obj(help='"MetricRelabelConfigs to apply to samples before ingestion."'), + metricRelabelings: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + '#oauth2':: d.obj(help='"OAuth2 for the URL. Only valid in Prometheus versions 2.27.0 and newer."'), + oauth2: { + '#clientId':: d.obj(help='"The secret or configmap containing the OAuth2 client id"'), + clientId: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { secret+: { optional: optional } } } }, + }, + }, + '#clientSecret':: d.obj(help='"The secret containing the OAuth2 client secret"'), + clientSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientSecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientSecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientSecret+: { optional: optional } } }, + }, + '#withEndpointParams':: d.fn(help='"Parameters to append to the token URL"', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParams(endpointParams): { oauth2+: { endpointParams: endpointParams } }, + '#withEndpointParamsMixin':: d.fn(help='"Parameters to append to the token URL"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParamsMixin(endpointParams): { oauth2+: { endpointParams+: endpointParams } }, + '#withScopes':: d.fn(help='"OAuth2 scopes used for the token request"', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { oauth2+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withScopesMixin':: d.fn(help='"OAuth2 scopes used for the token request"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { oauth2+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withTokenUrl':: d.fn(help='"The URL to fetch the token from"', args=[d.arg(name='tokenUrl', type=d.T.string)]), + withTokenUrl(tokenUrl): { oauth2+: { tokenUrl: tokenUrl } }, + }, + '#relabelings':: d.obj(help="\"RelabelConfigs to apply to samples before scraping. Prometheus Operator automatically adds relabelings for a few standard Kubernetes fields. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\""), + relabelings: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + '#tlsConfig':: d.obj(help='"TLS configuration to use when scraping the endpoint."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { secret+: { optional: optional } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { secret+: { optional: optional } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { keySecret+: { optional: optional } } }, + }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { tlsConfig+: { serverName: serverName } }, + }, + '#withEnableHttp2':: d.fn(help='"Whether to enable HTTP2."', args=[d.arg(name='enableHttp2', type=d.T.boolean)]), + withEnableHttp2(enableHttp2): { enableHttp2: enableHttp2 }, + '#withFilterRunning':: d.fn(help='"Drop pods that are not running. (Failed, Succeeded). Enabled by default. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#pod-phase"', args=[d.arg(name='filterRunning', type=d.T.boolean)]), + withFilterRunning(filterRunning): { filterRunning: filterRunning }, + '#withFollowRedirects':: d.fn(help='"FollowRedirects configures whether scrape requests follow HTTP 3xx redirects."', args=[d.arg(name='followRedirects', type=d.T.boolean)]), + withFollowRedirects(followRedirects): { followRedirects: followRedirects }, + '#withHonorLabels':: d.fn(help="\"HonorLabels chooses the metric's labels on collisions with target labels.\"", args=[d.arg(name='honorLabels', type=d.T.boolean)]), + withHonorLabels(honorLabels): { honorLabels: honorLabels }, + '#withHonorTimestamps':: d.fn(help='"HonorTimestamps controls whether Prometheus respects the timestamps present in scraped data."', args=[d.arg(name='honorTimestamps', type=d.T.boolean)]), + withHonorTimestamps(honorTimestamps): { honorTimestamps: honorTimestamps }, + '#withInterval':: d.fn(help="\"Interval at which metrics should be scraped If not specified Prometheus' global scrape interval is used.\"", args=[d.arg(name='interval', type=d.T.string)]), + withInterval(interval): { interval: interval }, + '#withMetricRelabelings':: d.fn(help='"MetricRelabelConfigs to apply to samples before ingestion."', args=[d.arg(name='metricRelabelings', type=d.T.array)]), + withMetricRelabelings(metricRelabelings): { metricRelabelings: if std.isArray(v=metricRelabelings) then metricRelabelings else [metricRelabelings] }, + '#withMetricRelabelingsMixin':: d.fn(help='"MetricRelabelConfigs to apply to samples before ingestion."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metricRelabelings', type=d.T.array)]), + withMetricRelabelingsMixin(metricRelabelings): { metricRelabelings+: if std.isArray(v=metricRelabelings) then metricRelabelings else [metricRelabelings] }, + '#withParams':: d.fn(help='"Optional HTTP URL parameters"', args=[d.arg(name='params', type=d.T.object)]), + withParams(params): { params: params }, + '#withParamsMixin':: d.fn(help='"Optional HTTP URL parameters"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='params', type=d.T.object)]), + withParamsMixin(params): { params+: params }, + '#withPath':: d.fn(help='"HTTP path to scrape for metrics. If empty, Prometheus uses the default value (e.g. `/metrics`)."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + '#withPort':: d.fn(help='"Name of the pod port this endpoint refers to. Mutually exclusive with targetPort."', args=[d.arg(name='port', type=d.T.string)]), + withPort(port): { port: port }, + '#withProxyUrl':: d.fn(help='"ProxyURL eg http://proxyserver:2195 Directs scrapes to proxy through this endpoint."', args=[d.arg(name='proxyUrl', type=d.T.string)]), + withProxyUrl(proxyUrl): { proxyUrl: proxyUrl }, + '#withRelabelings':: d.fn(help="\"RelabelConfigs to apply to samples before scraping. Prometheus Operator automatically adds relabelings for a few standard Kubernetes fields. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\"", args=[d.arg(name='relabelings', type=d.T.array)]), + withRelabelings(relabelings): { relabelings: if std.isArray(v=relabelings) then relabelings else [relabelings] }, + '#withRelabelingsMixin':: d.fn(help="\"RelabelConfigs to apply to samples before scraping. Prometheus Operator automatically adds relabelings for a few standard Kubernetes fields. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='relabelings', type=d.T.array)]), + withRelabelingsMixin(relabelings): { relabelings+: if std.isArray(v=relabelings) then relabelings else [relabelings] }, + '#withScheme':: d.fn(help='"HTTP scheme to use for scraping. `http` and `https` are the expected values unless you rewrite the `__scheme__` label via relabeling. If empty, Prometheus uses the default value `http`."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { scheme: scheme }, + '#withScrapeTimeout':: d.fn(help='"Timeout after which the scrape is ended If not specified, the Prometheus global scrape interval is used."', args=[d.arg(name='scrapeTimeout', type=d.T.string)]), + withScrapeTimeout(scrapeTimeout): { scrapeTimeout: scrapeTimeout }, + '#withTargetPort':: d.fn(help="\"Deprecated: Use 'port' instead.\"", args=[d.arg(name='targetPort', type=d.T.any)]), + withTargetPort(targetPort): { targetPort: targetPort }, + }, + '#selector':: d.obj(help='"Selector to select Pod objects."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { selector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { selector+: { matchLabels+: matchLabels } } }, + }, + '#withJobLabel':: d.fn(help='"The label to use to retrieve the job name from."', args=[d.arg(name='jobLabel', type=d.T.string)]), + withJobLabel(jobLabel): { spec+: { jobLabel: jobLabel } }, + '#withLabelLimit':: d.fn(help='"Per-scrape limit on number of labels that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelLimit', type=d.T.integer)]), + withLabelLimit(labelLimit): { spec+: { labelLimit: labelLimit } }, + '#withLabelNameLengthLimit':: d.fn(help='"Per-scrape limit on length of labels name that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelNameLengthLimit', type=d.T.integer)]), + withLabelNameLengthLimit(labelNameLengthLimit): { spec+: { labelNameLengthLimit: labelNameLengthLimit } }, + '#withLabelValueLengthLimit':: d.fn(help='"Per-scrape limit on length of labels value that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelValueLengthLimit', type=d.T.integer)]), + withLabelValueLengthLimit(labelValueLengthLimit): { spec+: { labelValueLengthLimit: labelValueLengthLimit } }, + '#withPodMetricsEndpoints':: d.fn(help='"A list of endpoints allowed as part of this PodMonitor."', args=[d.arg(name='podMetricsEndpoints', type=d.T.array)]), + withPodMetricsEndpoints(podMetricsEndpoints): { spec+: { podMetricsEndpoints: if std.isArray(v=podMetricsEndpoints) then podMetricsEndpoints else [podMetricsEndpoints] } }, + '#withPodMetricsEndpointsMixin':: d.fn(help='"A list of endpoints allowed as part of this PodMonitor."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podMetricsEndpoints', type=d.T.array)]), + withPodMetricsEndpointsMixin(podMetricsEndpoints): { spec+: { podMetricsEndpoints+: if std.isArray(v=podMetricsEndpoints) then podMetricsEndpoints else [podMetricsEndpoints] } }, + '#withPodTargetLabels':: d.fn(help='"PodTargetLabels transfers labels on the Kubernetes Pod onto the target."', args=[d.arg(name='podTargetLabels', type=d.T.array)]), + withPodTargetLabels(podTargetLabels): { spec+: { podTargetLabels: if std.isArray(v=podTargetLabels) then podTargetLabels else [podTargetLabels] } }, + '#withPodTargetLabelsMixin':: d.fn(help='"PodTargetLabels transfers labels on the Kubernetes Pod onto the target."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='podTargetLabels', type=d.T.array)]), + withPodTargetLabelsMixin(podTargetLabels): { spec+: { podTargetLabels+: if std.isArray(v=podTargetLabels) then podTargetLabels else [podTargetLabels] } }, + '#withSampleLimit':: d.fn(help='"SampleLimit defines per-scrape limit on number of scraped samples that will be accepted."', args=[d.arg(name='sampleLimit', type=d.T.integer)]), + withSampleLimit(sampleLimit): { spec+: { sampleLimit: sampleLimit } }, + '#withTargetLimit':: d.fn(help='"TargetLimit defines a limit on the number of scraped targets that will be accepted."', args=[d.arg(name='targetLimit', type=d.T.integer)]), + withTargetLimit(targetLimit): { spec+: { targetLimit: targetLimit } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/0.13/_gen/monitoring/v1/probe.libsonnet b/0.13/_gen/monitoring/v1/probe.libsonnet new file mode 100644 index 0000000..d52555a --- /dev/null +++ b/0.13/_gen/monitoring/v1/probe.libsonnet @@ -0,0 +1,346 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='probe', url='', help='"Probe defines monitoring for a set of static targets or ingresses."'), + '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Probe', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'monitoring.coreos.com/v1', + kind: 'Probe', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='"Specification of desired Ingress selection for target discovery by Prometheus."'), + spec: { + '#authorization':: d.obj(help='"Authorization section for this endpoint"'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { authorization+: { credentials+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { authorization+: { credentials+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { authorization+: { credentials+: { optional: optional } } } }, + }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { authorization+: { type: type } } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth allow an endpoint to authenticate over basic authentication. More info: https://prometheus.io/docs/operating/configuration/#endpoint"'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { basicAuth+: { password+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { basicAuth+: { password+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { basicAuth+: { password+: { optional: optional } } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { basicAuth+: { username+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { basicAuth+: { username+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { basicAuth+: { username+: { optional: optional } } } }, + }, + }, + '#bearerTokenSecret':: d.obj(help='"Secret to mount to read bearer token for scraping targets. The secret needs to be in the same namespace as the probe and accessible by the Prometheus Operator."'), + bearerTokenSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { bearerTokenSecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { bearerTokenSecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { bearerTokenSecret+: { optional: optional } } }, + }, + '#metricRelabelings':: d.obj(help='"MetricRelabelConfigs to apply to samples before ingestion."'), + metricRelabelings: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + '#oauth2':: d.obj(help='"OAuth2 for the URL. Only valid in Prometheus versions 2.27.0 and newer."'), + oauth2: { + '#clientId':: d.obj(help='"The secret or configmap containing the OAuth2 client id"'), + clientId: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { oauth2+: { clientId+: { configMap+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { oauth2+: { clientId+: { configMap+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { oauth2+: { clientId+: { configMap+: { optional: optional } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { oauth2+: { clientId+: { secret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { oauth2+: { clientId+: { secret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { oauth2+: { clientId+: { secret+: { optional: optional } } } } }, + }, + }, + '#clientSecret':: d.obj(help='"The secret containing the OAuth2 client secret"'), + clientSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { oauth2+: { clientSecret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { oauth2+: { clientSecret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { oauth2+: { clientSecret+: { optional: optional } } } }, + }, + '#withEndpointParams':: d.fn(help='"Parameters to append to the token URL"', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParams(endpointParams): { spec+: { oauth2+: { endpointParams: endpointParams } } }, + '#withEndpointParamsMixin':: d.fn(help='"Parameters to append to the token URL"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParamsMixin(endpointParams): { spec+: { oauth2+: { endpointParams+: endpointParams } } }, + '#withScopes':: d.fn(help='"OAuth2 scopes used for the token request"', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { spec+: { oauth2+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } } }, + '#withScopesMixin':: d.fn(help='"OAuth2 scopes used for the token request"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { spec+: { oauth2+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } } }, + '#withTokenUrl':: d.fn(help='"The URL to fetch the token from"', args=[d.arg(name='tokenUrl', type=d.T.string)]), + withTokenUrl(tokenUrl): { spec+: { oauth2+: { tokenUrl: tokenUrl } } }, + }, + '#prober':: d.obj(help='"Specification for the prober to use for probing targets. The prober.URL parameter is required. Targets cannot be probed if left empty."'), + prober: { + '#withPath':: d.fn(help='"Path to collect metrics from. Defaults to `/probe`."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { spec+: { prober+: { path: path } } }, + '#withProxyUrl':: d.fn(help='"Optional ProxyURL."', args=[d.arg(name='proxyUrl', type=d.T.string)]), + withProxyUrl(proxyUrl): { spec+: { prober+: { proxyUrl: proxyUrl } } }, + '#withScheme':: d.fn(help='"HTTP scheme to use for scraping. `http` and `https` are the expected values unless you rewrite the `__scheme__` label via relabeling. If empty, Prometheus uses the default value `http`."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { spec+: { prober+: { scheme: scheme } } }, + '#withUrl':: d.fn(help='"Mandatory URL of the prober."', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { spec+: { prober+: { url: url } } }, + }, + '#targets':: d.obj(help='"Targets defines a set of static or dynamically discovered targets to probe."'), + targets: { + '#ingress':: d.obj(help='"ingress defines the Ingress objects to probe and the relabeling configuration. If `staticConfig` is also defined, `staticConfig` takes precedence."'), + ingress: { + '#namespaceSelector':: d.obj(help='"From which namespaces to select Ingress objects."'), + namespaceSelector: { + '#withAny':: d.fn(help='"Boolean describing whether all namespaces are selected in contrast to a list restricting them."', args=[d.arg(name='any', type=d.T.boolean)]), + withAny(any): { spec+: { targets+: { ingress+: { namespaceSelector+: { any: any } } } } }, + '#withMatchNames':: d.fn(help='"List of namespace names to select from."', args=[d.arg(name='matchNames', type=d.T.array)]), + withMatchNames(matchNames): { spec+: { targets+: { ingress+: { namespaceSelector+: { matchNames: if std.isArray(v=matchNames) then matchNames else [matchNames] } } } } }, + '#withMatchNamesMixin':: d.fn(help='"List of namespace names to select from."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchNames', type=d.T.array)]), + withMatchNamesMixin(matchNames): { spec+: { targets+: { ingress+: { namespaceSelector+: { matchNames+: if std.isArray(v=matchNames) then matchNames else [matchNames] } } } } }, + }, + '#relabelingConfigs':: d.obj(help="\"RelabelConfigs to apply to the label set of the target before it gets scraped. The original ingress address is available via the `__tmp_prometheus_ingress_address` label. It can be used to customize the probed URL. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\""), + relabelingConfigs: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + '#selector':: d.obj(help='"Selector to select the Ingress objects."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { targets+: { ingress+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { targets+: { ingress+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { targets+: { ingress+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { targets+: { ingress+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#withRelabelingConfigs':: d.fn(help="\"RelabelConfigs to apply to the label set of the target before it gets scraped. The original ingress address is available via the `__tmp_prometheus_ingress_address` label. It can be used to customize the probed URL. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\"", args=[d.arg(name='relabelingConfigs', type=d.T.array)]), + withRelabelingConfigs(relabelingConfigs): { spec+: { targets+: { ingress+: { relabelingConfigs: if std.isArray(v=relabelingConfigs) then relabelingConfigs else [relabelingConfigs] } } } }, + '#withRelabelingConfigsMixin':: d.fn(help="\"RelabelConfigs to apply to the label set of the target before it gets scraped. The original ingress address is available via the `__tmp_prometheus_ingress_address` label. It can be used to customize the probed URL. The original scrape job's name is available via the `__tmp_prometheus_job_name` label. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='relabelingConfigs', type=d.T.array)]), + withRelabelingConfigsMixin(relabelingConfigs): { spec+: { targets+: { ingress+: { relabelingConfigs+: if std.isArray(v=relabelingConfigs) then relabelingConfigs else [relabelingConfigs] } } } }, + }, + '#staticConfig':: d.obj(help='"staticConfig defines the static list of targets to probe and the relabeling configuration. If `ingress` is also defined, `staticConfig` takes precedence. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#static_config."'), + staticConfig: { + '#relabelingConfigs':: d.obj(help='"RelabelConfigs to apply to the label set of the targets before it gets scraped. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config"'), + relabelingConfigs: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + '#withLabels':: d.fn(help='"Labels assigned to all metrics scraped from the targets."', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { targets+: { staticConfig+: { labels: labels } } } }, + '#withLabelsMixin':: d.fn(help='"Labels assigned to all metrics scraped from the targets."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { targets+: { staticConfig+: { labels+: labels } } } }, + '#withRelabelingConfigs':: d.fn(help='"RelabelConfigs to apply to the label set of the targets before it gets scraped. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config"', args=[d.arg(name='relabelingConfigs', type=d.T.array)]), + withRelabelingConfigs(relabelingConfigs): { spec+: { targets+: { staticConfig+: { relabelingConfigs: if std.isArray(v=relabelingConfigs) then relabelingConfigs else [relabelingConfigs] } } } }, + '#withRelabelingConfigsMixin':: d.fn(help='"RelabelConfigs to apply to the label set of the targets before it gets scraped. More info: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#relabel_config"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='relabelingConfigs', type=d.T.array)]), + withRelabelingConfigsMixin(relabelingConfigs): { spec+: { targets+: { staticConfig+: { relabelingConfigs+: if std.isArray(v=relabelingConfigs) then relabelingConfigs else [relabelingConfigs] } } } }, + '#withStatic':: d.fn(help='"The list of hosts to probe."', args=[d.arg(name='static', type=d.T.array)]), + withStatic(static): { spec+: { targets+: { staticConfig+: { static: if std.isArray(v=static) then static else [static] } } } }, + '#withStaticMixin':: d.fn(help='"The list of hosts to probe."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='static', type=d.T.array)]), + withStaticMixin(static): { spec+: { targets+: { staticConfig+: { static+: if std.isArray(v=static) then static else [static] } } } }, + }, + }, + '#tlsConfig':: d.obj(help='"TLS configuration to use when scraping the endpoint."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tlsConfig+: { ca+: { configMap+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tlsConfig+: { ca+: { configMap+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tlsConfig+: { ca+: { configMap+: { optional: optional } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tlsConfig+: { ca+: { secret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tlsConfig+: { ca+: { secret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tlsConfig+: { ca+: { secret+: { optional: optional } } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tlsConfig+: { cert+: { secret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tlsConfig+: { cert+: { secret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tlsConfig+: { keySecret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tlsConfig+: { keySecret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tlsConfig+: { keySecret+: { optional: optional } } } }, + }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { spec+: { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { spec+: { tlsConfig+: { serverName: serverName } } }, + }, + '#withInterval':: d.fn(help="\"Interval at which targets are probed using the configured prober. If not specified Prometheus' global scrape interval is used.\"", args=[d.arg(name='interval', type=d.T.string)]), + withInterval(interval): { spec+: { interval: interval } }, + '#withJobName':: d.fn(help='"The job name assigned to scraped metrics by default."', args=[d.arg(name='jobName', type=d.T.string)]), + withJobName(jobName): { spec+: { jobName: jobName } }, + '#withLabelLimit':: d.fn(help='"Per-scrape limit on number of labels that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelLimit', type=d.T.integer)]), + withLabelLimit(labelLimit): { spec+: { labelLimit: labelLimit } }, + '#withLabelNameLengthLimit':: d.fn(help='"Per-scrape limit on length of labels name that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelNameLengthLimit', type=d.T.integer)]), + withLabelNameLengthLimit(labelNameLengthLimit): { spec+: { labelNameLengthLimit: labelNameLengthLimit } }, + '#withLabelValueLengthLimit':: d.fn(help='"Per-scrape limit on length of labels value that will be accepted for a sample. Only valid in Prometheus versions 2.27.0 and newer."', args=[d.arg(name='labelValueLengthLimit', type=d.T.integer)]), + withLabelValueLengthLimit(labelValueLengthLimit): { spec+: { labelValueLengthLimit: labelValueLengthLimit } }, + '#withMetricRelabelings':: d.fn(help='"MetricRelabelConfigs to apply to samples before ingestion."', args=[d.arg(name='metricRelabelings', type=d.T.array)]), + withMetricRelabelings(metricRelabelings): { spec+: { metricRelabelings: if std.isArray(v=metricRelabelings) then metricRelabelings else [metricRelabelings] } }, + '#withMetricRelabelingsMixin':: d.fn(help='"MetricRelabelConfigs to apply to samples before ingestion."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metricRelabelings', type=d.T.array)]), + withMetricRelabelingsMixin(metricRelabelings): { spec+: { metricRelabelings+: if std.isArray(v=metricRelabelings) then metricRelabelings else [metricRelabelings] } }, + '#withModule':: d.fn(help='"The module to use for probing specifying how to probe the target. Example module configuring in the blackbox exporter: https://github.com/prometheus/blackbox_exporter/blob/master/example.yml"', args=[d.arg(name='module', type=d.T.string)]), + withModule(module): { spec+: { module: module } }, + '#withSampleLimit':: d.fn(help='"SampleLimit defines per-scrape limit on number of scraped samples that will be accepted."', args=[d.arg(name='sampleLimit', type=d.T.integer)]), + withSampleLimit(sampleLimit): { spec+: { sampleLimit: sampleLimit } }, + '#withScrapeTimeout':: d.fn(help='"Timeout for scraping metrics from the Prometheus exporter. If not specified, the Prometheus global scrape timeout is used."', args=[d.arg(name='scrapeTimeout', type=d.T.string)]), + withScrapeTimeout(scrapeTimeout): { spec+: { scrapeTimeout: scrapeTimeout } }, + '#withTargetLimit':: d.fn(help='"TargetLimit defines a limit on the number of scraped targets that will be accepted."', args=[d.arg(name='targetLimit', type=d.T.integer)]), + withTargetLimit(targetLimit): { spec+: { targetLimit: targetLimit } }, + }, + '#mixin': 'ignore', + mixin: self, +} diff --git a/0.13/_gen/monitoring/v1/prometheus.libsonnet b/0.13/_gen/monitoring/v1/prometheus.libsonnet new file mode 100644 index 0000000..c512a57 --- /dev/null +++ b/0.13/_gen/monitoring/v1/prometheus.libsonnet @@ -0,0 +1,3799 @@ +{ + local d = (import 'doc-util/main.libsonnet'), + '#':: d.pkg(name='prometheus', url='', help='"Prometheus defines a Prometheus deployment."'), + '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { metadata+: { annotations: annotations } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } }, + '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]), + withClusterName(clusterName): { metadata+: { clusterName: clusterName } }, + '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]), + withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } }, + '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]), + withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } }, + '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]), + withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } }, + '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]), + withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } }, + '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]), + withGenerateName(generateName): { metadata+: { generateName: generateName } }, + '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]), + withGeneration(generation): { metadata+: { generation: generation } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { metadata+: { labels: labels } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { metadata+: { labels+: labels } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { metadata+: { name: name } }, + '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { metadata+: { namespace: namespace } }, + '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]), + withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } }, + '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]), + withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } }, + '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]), + withSelfLink(selfLink): { metadata+: { selfLink: selfLink } }, + '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]), + withUid(uid): { metadata+: { uid: uid } }, + }, + '#new':: d.fn(help='new returns an instance of Prometheus', args=[d.arg(name='name', type=d.T.string)]), + new(name): { + apiVersion: 'monitoring.coreos.com/v1', + kind: 'Prometheus', + } + self.metadata.withName(name=name), + '#spec':: d.obj(help='"Specification of the desired behavior of the Prometheus cluster. More info: https://github.com/kubernetes/community/blob/master/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'), + spec: { + '#additionalAlertManagerConfigs':: d.obj(help='"AdditionalAlertManagerConfigs specifies a key of a Secret containing additional Prometheus Alertmanager configurations. The Alertmanager configurations are appended to the configuration generated by the Prometheus Operator. They must be formatted according to the official Prometheus documentation: \\n https://prometheus.io/docs/prometheus/latest/configuration/configuration/#alertmanager_config \\n The user is responsible for making sure that the configurations are valid \\n Note that using this feature may expose the possibility to break upgrades of Prometheus. It is advised to review Prometheus release notes to ensure that no incompatible AlertManager configs are going to break Prometheus after the upgrade."'), + additionalAlertManagerConfigs: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { additionalAlertManagerConfigs+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { additionalAlertManagerConfigs+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { additionalAlertManagerConfigs+: { optional: optional } } }, + }, + '#additionalAlertRelabelConfigs':: d.obj(help='"AdditionalAlertRelabelConfigs specifies a key of a Secret containing additional Prometheus alert relabel configurations. The alert relabel configurations are appended to the configuration generated by the Prometheus Operator. They must be formatted according to the official Prometheus documentation: \\n https://prometheus.io/docs/prometheus/latest/configuration/configuration/#alert_relabel_configs \\n The user is responsible for making sure that the configurations are valid \\n Note that using this feature may expose the possibility to break upgrades of Prometheus. It is advised to review Prometheus release notes to ensure that no incompatible alert relabel configs are going to break Prometheus after the upgrade."'), + additionalAlertRelabelConfigs: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { additionalAlertRelabelConfigs+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { additionalAlertRelabelConfigs+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { additionalAlertRelabelConfigs+: { optional: optional } } }, + }, + '#additionalArgs':: d.obj(help="\"AdditionalArgs allows setting additional arguments for the 'prometheus' container. \\n It is intended for e.g. activating hidden flags which are not supported by the dedicated configuration options yet. The arguments are passed as-is to the Prometheus container which may cause issues if they are invalid or not supported by the given Prometheus version. \\n In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged.\""), + additionalArgs: { + '#withName':: d.fn(help='"Name of the argument, e.g. \\"scrape.discovery-reload-interval\\"."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Argument value, e.g. 30s. Can be empty for name-only arguments (e.g. --storage.tsdb.no-lockfile)"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#additionalScrapeConfigs':: d.obj(help='"AdditionalScrapeConfigs allows specifying a key of a Secret containing additional Prometheus scrape configurations. Scrape configurations specified are appended to the configurations generated by the Prometheus Operator. Job configurations specified must have the form as specified in the official Prometheus documentation: https://prometheus.io/docs/prometheus/latest/configuration/configuration/#scrape_config. As scrape configs are appended, the user is responsible to make sure it is valid. Note that using this feature may expose the possibility to break upgrades of Prometheus. It is advised to review Prometheus release notes to ensure that no incompatible scrape configs are going to break Prometheus after the upgrade."'), + additionalScrapeConfigs: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { additionalScrapeConfigs+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { additionalScrapeConfigs+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { additionalScrapeConfigs+: { optional: optional } } }, + }, + '#affinity':: d.obj(help="\"Defines the Pods' affinity scheduling rules if specified.\""), + affinity: { + '#nodeAffinity':: d.obj(help='"Describes node affinity scheduling rules for the pod."'), + nodeAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#preference':: d.obj(help='"A node selector term, associated with the corresponding weight."'), + preference: { + '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""), + matchExpressions: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""), + matchFields: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { preference+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { preference+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { preference+: { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { preference+: { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] } }, + }, + '#withWeight':: d.fn(help='"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#nodeSelectorTerms':: d.obj(help='"Required. A list of node selector terms. The terms are ORed."'), + nodeSelectorTerms: { + '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""), + matchExpressions: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""), + matchFields: { + '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] }, + '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFields(matchFields): { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]), + withMatchFieldsMixin(matchFields): { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] }, + }, + '#withNodeSelectorTerms':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTerms(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + '#withNodeSelectorTermsMixin':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]), + withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAffinity':: d.obj(help='"Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s))."'), + podAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'), + podAffinityTerm: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } }, + }, + '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + '#podAntiAffinity':: d.obj(help='"Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s))."'), + podAntiAffinity: { + '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'), + preferredDuringSchedulingIgnoredDuringExecution: { + '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'), + podAffinityTerm: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } }, + }, + '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]), + withWeight(weight): { weight: weight }, + }, + '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'), + requiredDuringSchedulingIgnoredDuringExecution: { + '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""), + namespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } }, + }, + '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]), + withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] }, + '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + }, + '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]), + withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } }, + }, + }, + '#alerting':: d.obj(help='"Defines the settings related to Alertmanager."'), + alerting: { + '#alertmanagers':: d.obj(help='"AlertmanagerEndpoints Prometheus should fire alerts against."'), + alertmanagers: { + '#authorization':: d.obj(help='"Authorization section for Alertmanager. \\n Cannot be set at the same time as `basicAuth`, or `bearerTokenFile`."'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { authorization+: { credentials+: { optional: optional } } }, + }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { authorization+: { type: type } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth configuration for Alertmanager. \\n Cannot be set at the same time as `bearerTokenFile`, or `authorization`."'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { password+: { optional: optional } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { username+: { optional: optional } } }, + }, + }, + '#tlsConfig':: d.obj(help='"TLS Config to use for Alertmanager."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { secret+: { optional: optional } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { secret+: { optional: optional } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { keySecret+: { optional: optional } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { tlsConfig+: { caFile: caFile } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { tlsConfig+: { certFile: certFile } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { tlsConfig+: { keyFile: keyFile } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { tlsConfig+: { serverName: serverName } }, + }, + '#withApiVersion':: d.fn(help='"Version of the Alertmanager API that Prometheus uses to send alerts. It can be \\"v1\\" or \\"v2\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { apiVersion: apiVersion }, + '#withBearerTokenFile':: d.fn(help='"File to read bearer token for Alertmanager. \\n Cannot be set at the same time as `basicAuth`, or `authorization`. \\n *Deprecated: this will be removed in a future release. Prefer using `authorization`.*"', args=[d.arg(name='bearerTokenFile', type=d.T.string)]), + withBearerTokenFile(bearerTokenFile): { bearerTokenFile: bearerTokenFile }, + '#withEnableHttp2':: d.fn(help='"Whether to enable HTTP2."', args=[d.arg(name='enableHttp2', type=d.T.boolean)]), + withEnableHttp2(enableHttp2): { enableHttp2: enableHttp2 }, + '#withName':: d.fn(help='"Name of the Endpoints object in the namespace."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='"Namespace of the Endpoints object."', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withPathPrefix':: d.fn(help='"Prefix for the HTTP path alerts are pushed to."', args=[d.arg(name='pathPrefix', type=d.T.string)]), + withPathPrefix(pathPrefix): { pathPrefix: pathPrefix }, + '#withPort':: d.fn(help='"Port on which the Alertmanager API is exposed."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { port: port }, + '#withScheme':: d.fn(help='"Scheme to use when firing alerts."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { scheme: scheme }, + '#withTimeout':: d.fn(help='"Timeout is a per-target Alertmanager timeout when pushing alerts."', args=[d.arg(name='timeout', type=d.T.string)]), + withTimeout(timeout): { timeout: timeout }, + }, + '#withAlertmanagers':: d.fn(help='"AlertmanagerEndpoints Prometheus should fire alerts against."', args=[d.arg(name='alertmanagers', type=d.T.array)]), + withAlertmanagers(alertmanagers): { spec+: { alerting+: { alertmanagers: if std.isArray(v=alertmanagers) then alertmanagers else [alertmanagers] } } }, + '#withAlertmanagersMixin':: d.fn(help='"AlertmanagerEndpoints Prometheus should fire alerts against."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='alertmanagers', type=d.T.array)]), + withAlertmanagersMixin(alertmanagers): { spec+: { alerting+: { alertmanagers+: if std.isArray(v=alertmanagers) then alertmanagers else [alertmanagers] } } }, + }, + '#apiserverConfig':: d.obj(help="\"APIServerConfig allows specifying a host and auth methods to access the Kuberntees API server. If null, Prometheus is assumed to run inside of the cluster: it will discover the API servers automatically and use the Pod's CA certificate and bearer token file at /var/run/secrets/kubernetes.io/serviceaccount/.\""), + apiserverConfig: { + '#authorization':: d.obj(help='"Authorization section for the API server. \\n Cannot be set at the same time as `basicAuth`, `bearerToken`, or `bearerTokenFile`."'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { authorization+: { credentials+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { authorization+: { credentials+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { authorization+: { credentials+: { optional: optional } } } } }, + }, + '#withCredentialsFile':: d.fn(help='"File to read a secret from, mutually exclusive with `credentials`."', args=[d.arg(name='credentialsFile', type=d.T.string)]), + withCredentialsFile(credentialsFile): { spec+: { apiserverConfig+: { authorization+: { credentialsFile: credentialsFile } } } }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { apiserverConfig+: { authorization+: { type: type } } } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth configuration for the API server. \\n Cannot be set at the same time as `authorization`, `bearerToken`, or `bearerTokenFile`."'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { basicAuth+: { password+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { basicAuth+: { password+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { basicAuth+: { password+: { optional: optional } } } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { basicAuth+: { username+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { basicAuth+: { username+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { basicAuth+: { username+: { optional: optional } } } } }, + }, + }, + '#tlsConfig':: d.obj(help='"TLS Config to use for the API server."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { tlsConfig+: { ca+: { secret+: { optional: optional } } } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { apiserverConfig+: { tlsConfig+: { keySecret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { apiserverConfig+: { tlsConfig+: { keySecret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { apiserverConfig+: { tlsConfig+: { keySecret+: { optional: optional } } } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { spec+: { apiserverConfig+: { tlsConfig+: { caFile: caFile } } } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { spec+: { apiserverConfig+: { tlsConfig+: { certFile: certFile } } } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { spec+: { apiserverConfig+: { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } } } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { spec+: { apiserverConfig+: { tlsConfig+: { keyFile: keyFile } } } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { spec+: { apiserverConfig+: { tlsConfig+: { serverName: serverName } } } }, + }, + '#withBearerToken':: d.fn(help="\"*Warning: this field shouldn't be used because the token value appears in clear-text. Prefer using `authorization`.* \\n *Deprecated: this will be removed in a future release.*\"", args=[d.arg(name='bearerToken', type=d.T.string)]), + withBearerToken(bearerToken): { spec+: { apiserverConfig+: { bearerToken: bearerToken } } }, + '#withBearerTokenFile':: d.fn(help='"File to read bearer token for accessing apiserver. \\n Cannot be set at the same time as `basicAuth`, `authorization`, or `bearerToken`. \\n *Deprecated: this will be removed in a future release. Prefer using `authorization`.*"', args=[d.arg(name='bearerTokenFile', type=d.T.string)]), + withBearerTokenFile(bearerTokenFile): { spec+: { apiserverConfig+: { bearerTokenFile: bearerTokenFile } } }, + '#withHost':: d.fn(help='"Kubernetes API address consisting of a hostname or IP address followed by an optional port number."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { spec+: { apiserverConfig+: { host: host } } }, + }, + '#arbitraryFSAccessThroughSMs':: d.obj(help="\"When true, ServiceMonitor, PodMonitor and Probe object are forbidden to reference arbitrary files on the file system of the 'prometheus' container. When a ServiceMonitor's endpoint specifies a `bearerTokenFile` value (e.g. '/var/run/secrets/kubernetes.io/serviceaccount/token'), a malicious target can get access to the Prometheus service account's token in the Prometheus' scrape request. Setting `spec.arbitraryFSAccessThroughSM` to 'true' would prevent the attack. Users should instead provide the credentials using the `spec.bearerTokenSecret` field.\""), + arbitraryFSAccessThroughSMs: { + '#withDeny':: d.fn(help='', args=[d.arg(name='deny', type=d.T.boolean)]), + withDeny(deny): { spec+: { arbitraryFSAccessThroughSMs+: { deny: deny } } }, + }, + '#containers':: d.obj(help='"Containers allows injecting additional containers or modifying operator generated containers. This can be used to allow adding an authentication proxy to the Pods or to change the behavior of an operator generated container. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. \\n The names of containers managed by the operator are: * `prometheus` * `config-reloader` * `thanos-sidecar` \\n Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."'), + containers: { + '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'), + env: { + '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""), + valueFrom: { + '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'), + configMapKeyRef: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } }, + }, + '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } }, + }, + '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""), + secretKeyRef: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { secretKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } }, + }, + }, + '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'), + envFrom: { + '#configMapRef':: d.obj(help='"The ConfigMap to select from"'), + configMapRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapRef+: { optional: optional } }, + }, + '#secretRef':: d.obj(help='"The Secret to select from"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretRef+: { optional: optional } }, + }, + '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]), + withPrefix(prefix): { prefix: prefix }, + }, + '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'), + lifecycle: { + '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'), + postStart: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""), + preStop: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + livenessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { livenessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { livenessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'), + ports: { + '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]), + withContainerPort(containerPort): { containerPort: containerPort }, + '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]), + withHostPort(hostPort): { hostPort: hostPort }, + '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + }, + '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + readinessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { readinessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { readinessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'), + resizePolicy: { + '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]), + withResourceName(resourceName): { resourceName: resourceName }, + '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy }, + }, + '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'), + securityContext: { + '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'), + capabilities: { + '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""), + startupProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { startupProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { startupProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'), + volumeDevices: { + '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + }, + '#excludedFromEnforcement':: d.obj(help='"List of references to PodMonitor, ServiceMonitor, Probe and PrometheusRule objects to be excluded from enforcing a namespace label of origin. \\n It is only applicable if `spec.enforcedNamespaceLabel` set to true."'), + excludedFromEnforcement: { + '#withGroup':: d.fn(help='"Group of the referent. When not specified, it defaults to `monitoring.coreos.com`"', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { group: group }, + '#withName':: d.fn(help='"Name of the referent. When not set, all resources in the namespace are matched."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withNamespace':: d.fn(help='"Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/"', args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { namespace: namespace }, + '#withResource':: d.fn(help='"Resource of the referent."', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resource: resource }, + }, + '#exemplars':: d.obj(help='"Exemplars related settings that are runtime reloadable. It requires to enable the `exemplar-storage` feature flag to be effective."'), + exemplars: { + '#withMaxSize':: d.fn(help='"Maximum number of exemplars stored in memory for all series. \\n exemplar-storage itself must be enabled using the `spec.enableFeature` option for exemplars to be scraped in the first place. \\n If not set, Prometheus uses its default value. A value of zero or less than zero disables the storage."', args=[d.arg(name='maxSize', type=d.T.integer)]), + withMaxSize(maxSize): { spec+: { exemplars+: { maxSize: maxSize } } }, + }, + '#hostAliases':: d.obj(help="\"Optional list of hosts and IPs that will be injected into the Pod's hosts file if specified.\""), + hostAliases: { + '#withHostnames':: d.fn(help='"Hostnames for the above IP address."', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnames(hostnames): { hostnames: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withHostnamesMixin':: d.fn(help='"Hostnames for the above IP address."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostnames', type=d.T.array)]), + withHostnamesMixin(hostnames): { hostnames+: if std.isArray(v=hostnames) then hostnames else [hostnames] }, + '#withIp':: d.fn(help='"IP address of the host file entry."', args=[d.arg(name='ip', type=d.T.string)]), + withIp(ip): { ip: ip }, + }, + '#imagePullSecrets':: d.obj(help='"An optional list of references to Secrets in the same namespace to use for pulling images from registries. See http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"'), + imagePullSecrets: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#initContainers':: d.obj(help='"InitContainers allows injecting initContainers to the Pod definition. Those can be used to e.g. fetch secrets for injection into the Prometheus configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. \\n The names of init container name managed by the operator are: * `init-config-reloader`. \\n Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."'), + initContainers: { + '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'), + env: { + '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""), + valueFrom: { + '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'), + configMapKeyRef: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } }, + }, + '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } }, + }, + '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""), + secretKeyRef: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { valueFrom+: { secretKeyRef+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } }, + }, + }, + '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'), + envFrom: { + '#configMapRef':: d.obj(help='"The ConfigMap to select from"'), + configMapRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMapRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMapRef+: { optional: optional } }, + }, + '#secretRef':: d.obj(help='"The Secret to select from"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secretRef+: { name: name } }, + '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secretRef+: { optional: optional } }, + }, + '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]), + withPrefix(prefix): { prefix: prefix }, + }, + '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'), + lifecycle: { + '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'), + postStart: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } }, + }, + }, + '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""), + preStop: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } }, + }, + '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } }, + }, + }, + }, + '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + livenessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { livenessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { livenessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { livenessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'), + ports: { + '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]), + withContainerPort(containerPort): { containerPort: containerPort }, + '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]), + withHostIP(hostIP): { hostIP: hostIP }, + '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]), + withHostPort(hostPort): { hostPort: hostPort }, + '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]), + withProtocol(protocol): { protocol: protocol }, + }, + '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'), + readinessProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { readinessProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { readinessProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { readinessProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'), + resizePolicy: { + '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]), + withResourceName(resourceName): { resourceName: resourceName }, + '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]), + withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy }, + }, + '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { resources+: { limits: limits } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { resources+: { limits+: limits } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { resources+: { requests: requests } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { resources+: { requests+: requests } }, + }, + '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'), + securityContext: { + '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'), + capabilities: { + '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]), + withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } }, + '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]), + withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } }, + '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]), + withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } }, + '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]), + withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } }, + }, + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seLinuxOptions+: { type: type } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { securityContext+: { seccompProfile+: { type: type } } }, + }, + '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } }, + }, + '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]), + withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } }, + '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]), + withPrivileged(privileged): { securityContext+: { privileged: privileged } }, + '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]), + withProcMount(procMount): { securityContext+: { procMount: procMount } }, + '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]), + withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } }, + }, + '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""), + startupProbe: { + '#exec':: d.obj(help='"Exec specifies the action to take."'), + exec: { + '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } }, + '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } }, + }, + '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'), + grpc: { + '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]), + withPort(port): { startupProbe+: { grpc+: { port: port } } }, + '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]), + withService(service): { startupProbe+: { grpc+: { service: service } } }, + }, + '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'), + httpGet: { + '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'), + httpHeaders: { + '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { httpGet+: { host: host } } }, + '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]), + withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } }, + '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { startupProbe+: { httpGet+: { path: path } } }, + '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { httpGet+: { port: port } } }, + '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]), + withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } }, + }, + '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'), + tcpSocket: { + '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]), + withHost(host): { startupProbe+: { tcpSocket+: { host: host } } }, + '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]), + withPort(port): { startupProbe+: { tcpSocket+: { port: port } } }, + }, + '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]), + withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } }, + '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]), + withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } }, + '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]), + withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } }, + '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]), + withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } }, + '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]), + withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } }, + '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]), + withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } }, + }, + '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'), + volumeDevices: { + '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]), + withDevicePath(devicePath): { devicePath: devicePath }, + '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]), + withArgs(args): { args: if std.isArray(v=args) then args else [args] }, + '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]), + withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] }, + '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]), + withCommand(command): { command: if std.isArray(v=command) then command else [command] }, + '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]), + withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] }, + '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]), + withEnv(env): { env: if std.isArray(v=env) then env else [env] }, + '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]), + withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] }, + '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]), + withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] }, + '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { image: image }, + '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy }, + '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]), + withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] }, + '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]), + withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] }, + '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]), + withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] }, + '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]), + withStdin(stdin): { stdin: stdin }, + '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]), + withStdinOnce(stdinOnce): { stdinOnce: stdinOnce }, + '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]), + withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath }, + '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]), + withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy }, + '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]), + withTty(tty): { tty: tty }, + '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]), + withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] }, + '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] }, + '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]), + withWorkingDir(workingDir): { workingDir: workingDir }, + }, + '#podMetadata':: d.obj(help='"PodMetadata configures labels and annotations which are propagated to the Prometheus pods."'), + podMetadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { podMetadata+: { annotations: annotations } } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { podMetadata+: { annotations+: annotations } } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { podMetadata+: { labels: labels } } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { podMetadata+: { labels+: labels } } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { podMetadata+: { name: name } } }, + }, + '#podMonitorNamespaceSelector':: d.obj(help='"Namespaces to match for PodMonitors discovery. An empty label selector matches all namespaces. A null label selector matches the current namespace only."'), + podMonitorNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { podMonitorNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { podMonitorNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { podMonitorNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { podMonitorNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#podMonitorSelector':: d.obj(help="\"*Experimental* PodMonitors to be selected for target discovery. An empty label selector matches all objects. A null label selector matches no objects. \\n If `spec.serviceMonitorSelector`, `spec.podMonitorSelector`, `spec.probeSelector` and `spec.scrapeConfigSelector` are null, the Prometheus configuration is unmanaged. The Prometheus operator will ensure that the Prometheus configuration's Secret exists, but it is the responsibility of the user to provide the raw gzipped Prometheus configuration under the `prometheus.yaml.gz` key. This behavior is *deprecated* and will be removed in the next major version of the custom resource definition. It is recommended to use `spec.additionalScrapeConfigs` instead.\""), + podMonitorSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { podMonitorSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { podMonitorSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { podMonitorSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { podMonitorSelector+: { matchLabels+: matchLabels } } }, + }, + '#probeNamespaceSelector':: d.obj(help='"*Experimental* Namespaces to match for Probe discovery. An empty label selector matches all namespaces. A null label selector matches the current namespace only."'), + probeNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { probeNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { probeNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { probeNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { probeNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#probeSelector':: d.obj(help="\"*Experimental* Probes to be selected for target discovery. An empty label selector matches all objects. A null label selector matches no objects. \\n If `spec.serviceMonitorSelector`, `spec.podMonitorSelector`, `spec.probeSelector` and `spec.scrapeConfigSelector` are null, the Prometheus configuration is unmanaged. The Prometheus operator will ensure that the Prometheus configuration's Secret exists, but it is the responsibility of the user to provide the raw gzipped Prometheus configuration under the `prometheus.yaml.gz` key. This behavior is *deprecated* and will be removed in the next major version of the custom resource definition. It is recommended to use `spec.additionalScrapeConfigs` instead.\""), + probeSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { probeSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { probeSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { probeSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { probeSelector+: { matchLabels+: matchLabels } } }, + }, + '#prometheusRulesExcludedFromEnforce':: d.obj(help="\"Defines the list of PrometheusRule objects to which the namespace label enforcement doesn't apply. This is only relevant when `spec.enforcedNamespaceLabel` is set to true. *Deprecated: use `spec.excludedFromEnforcement` instead.*\""), + prometheusRulesExcludedFromEnforce: { + '#withRuleName':: d.fn(help='"Name of the excluded PrometheusRule object."', args=[d.arg(name='ruleName', type=d.T.string)]), + withRuleName(ruleName): { ruleName: ruleName }, + '#withRuleNamespace':: d.fn(help='"Namespace of the excluded PrometheusRule object."', args=[d.arg(name='ruleNamespace', type=d.T.string)]), + withRuleNamespace(ruleNamespace): { ruleNamespace: ruleNamespace }, + }, + '#query':: d.obj(help='"QuerySpec defines the configuration of the Promethus query service."'), + query: { + '#withLookbackDelta':: d.fn(help='"The delta difference allowed for retrieving metrics during expression evaluations."', args=[d.arg(name='lookbackDelta', type=d.T.string)]), + withLookbackDelta(lookbackDelta): { spec+: { query+: { lookbackDelta: lookbackDelta } } }, + '#withMaxConcurrency':: d.fn(help='"Number of concurrent queries that can be run at once."', args=[d.arg(name='maxConcurrency', type=d.T.integer)]), + withMaxConcurrency(maxConcurrency): { spec+: { query+: { maxConcurrency: maxConcurrency } } }, + '#withMaxSamples':: d.fn(help='"Maximum number of samples a single query can load into memory. Note that queries will fail if they would load more samples than this into memory, so this also limits the number of samples a query can return."', args=[d.arg(name='maxSamples', type=d.T.integer)]), + withMaxSamples(maxSamples): { spec+: { query+: { maxSamples: maxSamples } } }, + '#withTimeout':: d.fn(help='"Maximum time a query may take before being aborted."', args=[d.arg(name='timeout', type=d.T.string)]), + withTimeout(timeout): { spec+: { query+: { timeout: timeout } } }, + }, + '#remoteRead':: d.obj(help='"Defines the list of remote read configurations."'), + remoteRead: { + '#authorization':: d.obj(help='"Authorization section for the URL. \\n It requires Prometheus >= v2.26.0. \\n Cannot be set at the same time as `basicAuth`, or `oauth2`."'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { authorization+: { credentials+: { optional: optional } } }, + }, + '#withCredentialsFile':: d.fn(help='"File to read a secret from, mutually exclusive with `credentials`."', args=[d.arg(name='credentialsFile', type=d.T.string)]), + withCredentialsFile(credentialsFile): { authorization+: { credentialsFile: credentialsFile } }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { authorization+: { type: type } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth configuration for the URL. \\n Cannot be set at the same time as `authorization`, or `oauth2`."'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { password+: { optional: optional } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { username+: { optional: optional } } }, + }, + }, + '#oauth2':: d.obj(help='"OAuth2 configuration for the URL. \\n It requires Prometheus >= v2.27.0. \\n Cannot be set at the same time as `authorization`, or `basicAuth`."'), + oauth2: { + '#clientId':: d.obj(help='"The secret or configmap containing the OAuth2 client id"'), + clientId: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { secret+: { optional: optional } } } }, + }, + }, + '#clientSecret':: d.obj(help='"The secret containing the OAuth2 client secret"'), + clientSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientSecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientSecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientSecret+: { optional: optional } } }, + }, + '#withEndpointParams':: d.fn(help='"Parameters to append to the token URL"', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParams(endpointParams): { oauth2+: { endpointParams: endpointParams } }, + '#withEndpointParamsMixin':: d.fn(help='"Parameters to append to the token URL"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParamsMixin(endpointParams): { oauth2+: { endpointParams+: endpointParams } }, + '#withScopes':: d.fn(help='"OAuth2 scopes used for the token request"', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { oauth2+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withScopesMixin':: d.fn(help='"OAuth2 scopes used for the token request"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { oauth2+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withTokenUrl':: d.fn(help='"The URL to fetch the token from"', args=[d.arg(name='tokenUrl', type=d.T.string)]), + withTokenUrl(tokenUrl): { oauth2+: { tokenUrl: tokenUrl } }, + }, + '#tlsConfig':: d.obj(help='"TLS Config to use for the URL."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { secret+: { optional: optional } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { secret+: { optional: optional } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { keySecret+: { optional: optional } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { tlsConfig+: { caFile: caFile } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { tlsConfig+: { certFile: certFile } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { tlsConfig+: { keyFile: keyFile } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { tlsConfig+: { serverName: serverName } }, + }, + '#withBearerToken':: d.fn(help="\"*Warning: this field shouldn't be used because the token value appears in clear-text. Prefer using `authorization`.* \\n *Deprecated: this will be removed in a future release.*\"", args=[d.arg(name='bearerToken', type=d.T.string)]), + withBearerToken(bearerToken): { bearerToken: bearerToken }, + '#withBearerTokenFile':: d.fn(help='"File from which to read the bearer token for the URL. \\n *Deprecated: this will be removed in a future release. Prefer using `authorization`.*"', args=[d.arg(name='bearerTokenFile', type=d.T.string)]), + withBearerTokenFile(bearerTokenFile): { bearerTokenFile: bearerTokenFile }, + '#withFilterExternalLabels':: d.fn(help='"Whether to use the external labels as selectors for the remote read endpoint. \\n It requires Prometheus >= v2.34.0."', args=[d.arg(name='filterExternalLabels', type=d.T.boolean)]), + withFilterExternalLabels(filterExternalLabels): { filterExternalLabels: filterExternalLabels }, + '#withFollowRedirects':: d.fn(help='"Configure whether HTTP requests follow HTTP 3xx redirects. \\n It requires Prometheus >= v2.26.0."', args=[d.arg(name='followRedirects', type=d.T.boolean)]), + withFollowRedirects(followRedirects): { followRedirects: followRedirects }, + '#withHeaders':: d.fn(help="\"Custom HTTP headers to be sent along with each remote read request. Be aware that headers that are set by Prometheus itself can't be overwritten. Only valid in Prometheus versions 2.26.0 and newer.\"", args=[d.arg(name='headers', type=d.T.object)]), + withHeaders(headers): { headers: headers }, + '#withHeadersMixin':: d.fn(help="\"Custom HTTP headers to be sent along with each remote read request. Be aware that headers that are set by Prometheus itself can't be overwritten. Only valid in Prometheus versions 2.26.0 and newer.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='headers', type=d.T.object)]), + withHeadersMixin(headers): { headers+: headers }, + '#withName':: d.fn(help='"The name of the remote read queue, it must be unique if specified. The name is used in metrics and logging in order to differentiate read configurations. \\n It requires Prometheus >= v2.15.0."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProxyUrl':: d.fn(help='"Optional ProxyURL."', args=[d.arg(name='proxyUrl', type=d.T.string)]), + withProxyUrl(proxyUrl): { proxyUrl: proxyUrl }, + '#withReadRecent':: d.fn(help='"Whether reads should be made for queries for time ranges that the local storage should have complete data for."', args=[d.arg(name='readRecent', type=d.T.boolean)]), + withReadRecent(readRecent): { readRecent: readRecent }, + '#withRemoteTimeout':: d.fn(help='"Timeout for requests to the remote read endpoint."', args=[d.arg(name='remoteTimeout', type=d.T.string)]), + withRemoteTimeout(remoteTimeout): { remoteTimeout: remoteTimeout }, + '#withRequiredMatchers':: d.fn(help='"An optional list of equality matchers which have to be present in a selector to query the remote read endpoint."', args=[d.arg(name='requiredMatchers', type=d.T.object)]), + withRequiredMatchers(requiredMatchers): { requiredMatchers: requiredMatchers }, + '#withRequiredMatchersMixin':: d.fn(help='"An optional list of equality matchers which have to be present in a selector to query the remote read endpoint."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredMatchers', type=d.T.object)]), + withRequiredMatchersMixin(requiredMatchers): { requiredMatchers+: requiredMatchers }, + '#withUrl':: d.fn(help='"The URL of the endpoint to query from."', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { url: url }, + }, + '#remoteWrite':: d.obj(help='"Defines the list of remote write configurations."'), + remoteWrite: { + '#authorization':: d.obj(help='"Authorization section for the URL. \\n It requires Prometheus >= v2.26.0. \\n Cannot be set at the same time as `sigv4`, `basicAuth`, or `oauth2`."'), + authorization: { + '#credentials':: d.obj(help='"Selects a key of a Secret in the namespace that contains the credentials for authentication."'), + credentials: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { authorization+: { credentials+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { authorization+: { credentials+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { authorization+: { credentials+: { optional: optional } } }, + }, + '#withCredentialsFile':: d.fn(help='"File to read a secret from, mutually exclusive with `credentials`."', args=[d.arg(name='credentialsFile', type=d.T.string)]), + withCredentialsFile(credentialsFile): { authorization+: { credentialsFile: credentialsFile } }, + '#withType':: d.fn(help='"Defines the authentication type. The value is case-insensitive. \\n \\"Basic\\" is not a supported value. \\n Default: \\"Bearer\\', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { authorization+: { type: type } }, + }, + '#basicAuth':: d.obj(help='"BasicAuth configuration for the URL. \\n Cannot be set at the same time as `sigv4`, `authorization`, or `oauth2`."'), + basicAuth: { + '#password':: d.obj(help='"The secret in the service monitor namespace that contains the password for authentication."'), + password: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { password+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { password+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { password+: { optional: optional } } }, + }, + '#username':: d.obj(help='"The secret in the service monitor namespace that contains the username for authentication."'), + username: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { basicAuth+: { username+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { basicAuth+: { username+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { basicAuth+: { username+: { optional: optional } } }, + }, + }, + '#metadataConfig':: d.obj(help='"MetadataConfig configures the sending of series metadata to the remote storage."'), + metadataConfig: { + '#withSend':: d.fn(help='"Defines whether metric metadata is sent to the remote storage or not."', args=[d.arg(name='send', type=d.T.boolean)]), + withSend(send): { metadataConfig+: { send: send } }, + '#withSendInterval':: d.fn(help='"Defines how frequently metric metadata is sent to the remote storage."', args=[d.arg(name='sendInterval', type=d.T.string)]), + withSendInterval(sendInterval): { metadataConfig+: { sendInterval: sendInterval } }, + }, + '#oauth2':: d.obj(help='"OAuth2 configuration for the URL. \\n It requires Prometheus >= v2.27.0. \\n Cannot be set at the same time as `sigv4`, `authorization`, or `basicAuth`."'), + oauth2: { + '#clientId':: d.obj(help='"The secret or configmap containing the OAuth2 client id"'), + clientId: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientId+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientId+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientId+: { secret+: { optional: optional } } } }, + }, + }, + '#clientSecret':: d.obj(help='"The secret containing the OAuth2 client secret"'), + clientSecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { oauth2+: { clientSecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { oauth2+: { clientSecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { oauth2+: { clientSecret+: { optional: optional } } }, + }, + '#withEndpointParams':: d.fn(help='"Parameters to append to the token URL"', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParams(endpointParams): { oauth2+: { endpointParams: endpointParams } }, + '#withEndpointParamsMixin':: d.fn(help='"Parameters to append to the token URL"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='endpointParams', type=d.T.object)]), + withEndpointParamsMixin(endpointParams): { oauth2+: { endpointParams+: endpointParams } }, + '#withScopes':: d.fn(help='"OAuth2 scopes used for the token request"', args=[d.arg(name='scopes', type=d.T.array)]), + withScopes(scopes): { oauth2+: { scopes: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withScopesMixin':: d.fn(help='"OAuth2 scopes used for the token request"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='scopes', type=d.T.array)]), + withScopesMixin(scopes): { oauth2+: { scopes+: if std.isArray(v=scopes) then scopes else [scopes] } }, + '#withTokenUrl':: d.fn(help='"The URL to fetch the token from"', args=[d.arg(name='tokenUrl', type=d.T.string)]), + withTokenUrl(tokenUrl): { oauth2+: { tokenUrl: tokenUrl } }, + }, + '#queueConfig':: d.obj(help='"QueueConfig allows tuning of the remote write queue parameters."'), + queueConfig: { + '#withBatchSendDeadline':: d.fn(help='"BatchSendDeadline is the maximum time a sample will wait in buffer."', args=[d.arg(name='batchSendDeadline', type=d.T.string)]), + withBatchSendDeadline(batchSendDeadline): { queueConfig+: { batchSendDeadline: batchSendDeadline } }, + '#withCapacity':: d.fn(help='"Capacity is the number of samples to buffer per shard before we start dropping them."', args=[d.arg(name='capacity', type=d.T.integer)]), + withCapacity(capacity): { queueConfig+: { capacity: capacity } }, + '#withMaxBackoff':: d.fn(help='"MaxBackoff is the maximum retry delay."', args=[d.arg(name='maxBackoff', type=d.T.string)]), + withMaxBackoff(maxBackoff): { queueConfig+: { maxBackoff: maxBackoff } }, + '#withMaxRetries':: d.fn(help='"MaxRetries is the maximum number of times to retry a batch on recoverable errors."', args=[d.arg(name='maxRetries', type=d.T.integer)]), + withMaxRetries(maxRetries): { queueConfig+: { maxRetries: maxRetries } }, + '#withMaxSamplesPerSend':: d.fn(help='"MaxSamplesPerSend is the maximum number of samples per send."', args=[d.arg(name='maxSamplesPerSend', type=d.T.integer)]), + withMaxSamplesPerSend(maxSamplesPerSend): { queueConfig+: { maxSamplesPerSend: maxSamplesPerSend } }, + '#withMaxShards':: d.fn(help='"MaxShards is the maximum number of shards, i.e. amount of concurrency."', args=[d.arg(name='maxShards', type=d.T.integer)]), + withMaxShards(maxShards): { queueConfig+: { maxShards: maxShards } }, + '#withMinBackoff':: d.fn(help='"MinBackoff is the initial retry delay. Gets doubled for every retry."', args=[d.arg(name='minBackoff', type=d.T.string)]), + withMinBackoff(minBackoff): { queueConfig+: { minBackoff: minBackoff } }, + '#withMinShards':: d.fn(help='"MinShards is the minimum number of shards, i.e. amount of concurrency."', args=[d.arg(name='minShards', type=d.T.integer)]), + withMinShards(minShards): { queueConfig+: { minShards: minShards } }, + '#withRetryOnRateLimit':: d.fn(help='"Retry upon receiving a 429 status code from the remote-write storage. This is experimental feature and might change in the future."', args=[d.arg(name='retryOnRateLimit', type=d.T.boolean)]), + withRetryOnRateLimit(retryOnRateLimit): { queueConfig+: { retryOnRateLimit: retryOnRateLimit } }, + }, + '#sigv4':: d.obj(help="\"Sigv4 allows to configures AWS's Signature Verification 4 for the URL. \\n It requires Prometheus \u003e= v2.26.0. \\n Cannot be set at the same time as `authorization`, `basicAuth`, or `oauth2`.\""), + sigv4: { + '#accessKey':: d.obj(help='"AccessKey is the AWS API key. If not specified, the environment variable `AWS_ACCESS_KEY_ID` is used."'), + accessKey: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { sigv4+: { accessKey+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { sigv4+: { accessKey+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { sigv4+: { accessKey+: { optional: optional } } }, + }, + '#secretKey':: d.obj(help='"SecretKey is the AWS API secret. If not specified, the environment variable `AWS_SECRET_ACCESS_KEY` is used."'), + secretKey: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { sigv4+: { secretKey+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { sigv4+: { secretKey+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { sigv4+: { secretKey+: { optional: optional } } }, + }, + '#withProfile':: d.fn(help='"Profile is the named AWS profile used to authenticate."', args=[d.arg(name='profile', type=d.T.string)]), + withProfile(profile): { sigv4+: { profile: profile } }, + '#withRegion':: d.fn(help='"Region is the AWS region. If blank, the region from the default credentials chain used."', args=[d.arg(name='region', type=d.T.string)]), + withRegion(region): { sigv4+: { region: region } }, + '#withRoleArn':: d.fn(help='"RoleArn is the named AWS profile used to authenticate."', args=[d.arg(name='roleArn', type=d.T.string)]), + withRoleArn(roleArn): { sigv4+: { roleArn: roleArn } }, + }, + '#tlsConfig':: d.obj(help='"TLS Config to use for the URL."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { ca+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { ca+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { ca+: { secret+: { optional: optional } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { configMap+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { configMap+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { configMap+: { optional: optional } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { cert+: { secret+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { cert+: { secret+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { cert+: { secret+: { optional: optional } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { tlsConfig+: { keySecret+: { key: key } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { tlsConfig+: { keySecret+: { name: name } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { tlsConfig+: { keySecret+: { optional: optional } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { tlsConfig+: { caFile: caFile } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { tlsConfig+: { certFile: certFile } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { tlsConfig+: { keyFile: keyFile } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { tlsConfig+: { serverName: serverName } }, + }, + '#withBearerToken':: d.fn(help="\"*Warning: this field shouldn't be used because the token value appears in clear-text. Prefer using `authorization`.* \\n *Deprecated: this will be removed in a future release.*\"", args=[d.arg(name='bearerToken', type=d.T.string)]), + withBearerToken(bearerToken): { bearerToken: bearerToken }, + '#withBearerTokenFile':: d.fn(help='"File from which to read bearer token for the URL. \\n *Deprecated: this will be removed in a future release. Prefer using `authorization`.*"', args=[d.arg(name='bearerTokenFile', type=d.T.string)]), + withBearerTokenFile(bearerTokenFile): { bearerTokenFile: bearerTokenFile }, + '#withHeaders':: d.fn(help="\"Custom HTTP headers to be sent along with each remote write request. Be aware that headers that are set by Prometheus itself can't be overwritten. \\n It requires Prometheus \u003e= v2.25.0.\"", args=[d.arg(name='headers', type=d.T.object)]), + withHeaders(headers): { headers: headers }, + '#withHeadersMixin':: d.fn(help="\"Custom HTTP headers to be sent along with each remote write request. Be aware that headers that are set by Prometheus itself can't be overwritten. \\n It requires Prometheus \u003e= v2.25.0.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='headers', type=d.T.object)]), + withHeadersMixin(headers): { headers+: headers }, + '#withName':: d.fn(help='"The name of the remote write queue, it must be unique if specified. The name is used in metrics and logging in order to differentiate queues. \\n It requires Prometheus >= v2.15.0."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withProxyUrl':: d.fn(help='"Optional ProxyURL."', args=[d.arg(name='proxyUrl', type=d.T.string)]), + withProxyUrl(proxyUrl): { proxyUrl: proxyUrl }, + '#withRemoteTimeout':: d.fn(help='"Timeout for requests to the remote write endpoint."', args=[d.arg(name='remoteTimeout', type=d.T.string)]), + withRemoteTimeout(remoteTimeout): { remoteTimeout: remoteTimeout }, + '#withSendExemplars':: d.fn(help='"Enables sending of exemplars over remote write. Note that exemplar-storage itself must be enabled using the `spec.enableFeature` option for exemplars to be scraped in the first place. \\n It requires Prometheus >= v2.27.0."', args=[d.arg(name='sendExemplars', type=d.T.boolean)]), + withSendExemplars(sendExemplars): { sendExemplars: sendExemplars }, + '#withSendNativeHistograms':: d.fn(help='"Enables sending of native histograms, also known as sparse histograms over remote write. \\n It requires Prometheus >= v2.40.0."', args=[d.arg(name='sendNativeHistograms', type=d.T.boolean)]), + withSendNativeHistograms(sendNativeHistograms): { sendNativeHistograms: sendNativeHistograms }, + '#withUrl':: d.fn(help='"The URL of the endpoint to send samples to."', args=[d.arg(name='url', type=d.T.string)]), + withUrl(url): { url: url }, + '#withWriteRelabelConfigs':: d.fn(help='"The list of remote write relabel configurations."', args=[d.arg(name='writeRelabelConfigs', type=d.T.array)]), + withWriteRelabelConfigs(writeRelabelConfigs): { writeRelabelConfigs: if std.isArray(v=writeRelabelConfigs) then writeRelabelConfigs else [writeRelabelConfigs] }, + '#withWriteRelabelConfigsMixin':: d.fn(help='"The list of remote write relabel configurations."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='writeRelabelConfigs', type=d.T.array)]), + withWriteRelabelConfigsMixin(writeRelabelConfigs): { writeRelabelConfigs+: if std.isArray(v=writeRelabelConfigs) then writeRelabelConfigs else [writeRelabelConfigs] }, + '#writeRelabelConfigs':: d.obj(help='"The list of remote write relabel configurations."'), + writeRelabelConfigs: { + '#withAction':: d.fn(help='"Action to perform based on the regex matching. \\n `Uppercase` and `Lowercase` actions require Prometheus >= v2.36.0. `DropEqual` and `KeepEqual` actions require Prometheus >= v2.41.0. \\n Default: \\"Replace\\', args=[d.arg(name='action', type=d.T.string)]), + withAction(action): { action: action }, + '#withModulus':: d.fn(help='"Modulus to take of the hash of the source label values. \\n Only applicable when the action is `HashMod`."', args=[d.arg(name='modulus', type=d.T.integer)]), + withModulus(modulus): { modulus: modulus }, + '#withRegex':: d.fn(help='"Regular expression against which the extracted value is matched."', args=[d.arg(name='regex', type=d.T.string)]), + withRegex(regex): { regex: regex }, + '#withReplacement':: d.fn(help='"Replacement value against which a Replace action is performed if the regular expression matches. \\n Regex capture groups are available."', args=[d.arg(name='replacement', type=d.T.string)]), + withReplacement(replacement): { replacement: replacement }, + '#withSeparator':: d.fn(help='"Separator is the string between concatenated SourceLabels."', args=[d.arg(name='separator', type=d.T.string)]), + withSeparator(separator): { separator: separator }, + '#withSourceLabels':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabels(sourceLabels): { sourceLabels: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withSourceLabelsMixin':: d.fn(help='"The source labels select values from existing labels. Their content is concatenated using the configured Separator and matched against the configured regular expression."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sourceLabels', type=d.T.array)]), + withSourceLabelsMixin(sourceLabels): { sourceLabels+: if std.isArray(v=sourceLabels) then sourceLabels else [sourceLabels] }, + '#withTargetLabel':: d.fn(help='"Label to which the resulting string is written in a replacement. \\n It is mandatory for `Replace`, `HashMod`, `Lowercase`, `Uppercase`, `KeepEqual` and `DropEqual` actions. \\n Regex capture groups are available."', args=[d.arg(name='targetLabel', type=d.T.string)]), + withTargetLabel(targetLabel): { targetLabel: targetLabel }, + }, + }, + '#resources':: d.obj(help="\"Defines the resources requests and limits of the 'prometheus' container.\""), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { resources+: { limits: limits } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { resources+: { limits+: limits } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { resources+: { requests: requests } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { resources+: { requests+: requests } } }, + }, + '#ruleNamespaceSelector':: d.obj(help='"Namespaces to match for PrometheusRule discovery. An empty label selector matches all namespaces. A null label selector matches the current namespace only."'), + ruleNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { ruleNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { ruleNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { ruleNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { ruleNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#ruleSelector':: d.obj(help='"PrometheusRule objects to be selected for rule evaluation. An empty label selector matches all objects. A null label selector matches no objects."'), + ruleSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { ruleSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { ruleSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { ruleSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { ruleSelector+: { matchLabels+: matchLabels } } }, + }, + '#rules':: d.obj(help="\"Defines the configuration of the Prometheus rules' engine.\""), + rules: { + '#alert':: d.obj(help="\"Defines the parameters of the Prometheus rules' engine. \\n Any update to these parameters trigger a restart of the pods.\""), + alert: { + '#withForGracePeriod':: d.fn(help="\"Minimum duration between alert and restored 'for' state. \\n This is maintained only for alerts with a configured 'for' time greater than the grace period.\"", args=[d.arg(name='forGracePeriod', type=d.T.string)]), + withForGracePeriod(forGracePeriod): { spec+: { rules+: { alert+: { forGracePeriod: forGracePeriod } } } }, + '#withForOutageTolerance':: d.fn(help="\"Max time to tolerate prometheus outage for restoring 'for' state of alert.\"", args=[d.arg(name='forOutageTolerance', type=d.T.string)]), + withForOutageTolerance(forOutageTolerance): { spec+: { rules+: { alert+: { forOutageTolerance: forOutageTolerance } } } }, + '#withResendDelay':: d.fn(help='"Minimum amount of time to wait before resending an alert to Alertmanager."', args=[d.arg(name='resendDelay', type=d.T.string)]), + withResendDelay(resendDelay): { spec+: { rules+: { alert+: { resendDelay: resendDelay } } } }, + }, + }, + '#scrapeConfigNamespaceSelector':: d.obj(help='"Namespaces to match for ScrapeConfig discovery. An empty label selector matches all namespaces. A null label selector matches the current current namespace only."'), + scrapeConfigNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { scrapeConfigNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { scrapeConfigNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { scrapeConfigNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { scrapeConfigNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#scrapeConfigSelector':: d.obj(help="\"*Experimental* ScrapeConfigs to be selected for target discovery. An empty label selector matches all objects. A null label selector matches no objects. \\n If `spec.serviceMonitorSelector`, `spec.podMonitorSelector`, `spec.probeSelector` and `spec.scrapeConfigSelector` are null, the Prometheus configuration is unmanaged. The Prometheus operator will ensure that the Prometheus configuration's Secret exists, but it is the responsibility of the user to provide the raw gzipped Prometheus configuration under the `prometheus.yaml.gz` key. This behavior is *deprecated* and will be removed in the next major version of the custom resource definition. It is recommended to use `spec.additionalScrapeConfigs` instead.\""), + scrapeConfigSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { scrapeConfigSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { scrapeConfigSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { scrapeConfigSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { scrapeConfigSelector+: { matchLabels+: matchLabels } } }, + }, + '#securityContext':: d.obj(help='"SecurityContext holds pod-level security attributes and common container settings. This defaults to the default PodSecurityContext."'), + securityContext: { + '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."'), + seLinuxOptions: { + '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]), + withLevel(level): { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } }, + '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]), + withRole(role): { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } }, + '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } }, + '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } }, + }, + '#seccompProfile':: d.obj(help='"The seccomp options to use by the containers in this pod. Note that this field cannot be set when spec.os.name is windows."'), + seccompProfile: { + '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must only be set if type is \\\"Localhost\\\".\"", args=[d.arg(name='localhostProfile', type=d.T.string)]), + withLocalhostProfile(localhostProfile): { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } }, + '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { spec+: { securityContext+: { seccompProfile+: { type: type } } } }, + }, + '#sysctls':: d.obj(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."'), + sysctls: { + '#withName':: d.fn(help='"Name of a property to set"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Value of a property to set"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#windowsOptions':: d.obj(help="\"The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux.\""), + windowsOptions: { + '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]), + withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } }, + '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]), + withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } }, + '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. This field is alpha-level and will only be honored by components that enable the WindowsHostProcessContainers feature flag. Setting this field without the feature flag will result in errors when validating the Pod. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]), + withHostProcess(hostProcess): { spec+: { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } } }, + '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]), + withRunAsUserName(runAsUserName): { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } }, + }, + '#withFsGroup':: d.fn(help="\"A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: \\n 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- \\n If unset, the Kubelet will not modify the ownership and permissions of any volume. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='fsGroup', type=d.T.integer)]), + withFsGroup(fsGroup): { spec+: { securityContext+: { fsGroup: fsGroup } } }, + '#withFsGroupChangePolicy':: d.fn(help='"fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are \\"OnRootMismatch\\" and \\"Always\\". If not specified, \\"Always\\" is used. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]), + withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } }, + '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]), + withRunAsGroup(runAsGroup): { spec+: { securityContext+: { runAsGroup: runAsGroup } } }, + '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]), + withRunAsNonRoot(runAsNonRoot): { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } }, + '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]), + withRunAsUser(runAsUser): { spec+: { securityContext+: { runAsUser: runAsUser } } }, + '#withSupplementalGroups':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroups(supplementalGroups): { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSupplementalGroupsMixin':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]), + withSupplementalGroupsMixin(supplementalGroups): { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } }, + '#withSysctls':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctls(sysctls): { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + '#withSysctlsMixin':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]), + withSysctlsMixin(sysctls): { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } }, + }, + '#serviceMonitorNamespaceSelector':: d.obj(help='"Namespaces to match for ServicedMonitors discovery. An empty label selector matches all namespaces. A null label selector matches the current namespace only."'), + serviceMonitorNamespaceSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { serviceMonitorNamespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { serviceMonitorNamespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { serviceMonitorNamespaceSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { serviceMonitorNamespaceSelector+: { matchLabels+: matchLabels } } }, + }, + '#serviceMonitorSelector':: d.obj(help="\"ServiceMonitors to be selected for target discovery. An empty label selector matches all objects. A null label selector matches no objects. \\n If `spec.serviceMonitorSelector`, `spec.podMonitorSelector`, `spec.probeSelector` and `spec.scrapeConfigSelector` are null, the Prometheus configuration is unmanaged. The Prometheus operator will ensure that the Prometheus configuration's Secret exists, but it is the responsibility of the user to provide the raw gzipped Prometheus configuration under the `prometheus.yaml.gz` key. This behavior is *deprecated* and will be removed in the next major version of the custom resource definition. It is recommended to use `spec.additionalScrapeConfigs` instead.\""), + serviceMonitorSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { serviceMonitorSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { serviceMonitorSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { serviceMonitorSelector+: { matchLabels: matchLabels } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { serviceMonitorSelector+: { matchLabels+: matchLabels } } }, + }, + '#storage':: d.obj(help='"Storage defines the storage used by Prometheus."'), + storage: { + '#emptyDir':: d.obj(help='"EmptyDirVolumeSource to be used by the StatefulSet. If specified, it takes precedence over `ephemeral` and `volumeClaimTemplate`. More info: https://kubernetes.io/docs/concepts/storage/volumes/#emptydir"'), + emptyDir: { + '#withMedium':: d.fn(help="\"medium represents what type of storage medium should back this directory. The default is \\\"\\\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\"", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { spec+: { storage+: { emptyDir+: { medium: medium } } } }, + '#withSizeLimit':: d.fn(help='"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"', args=[d.arg(name='sizeLimit', type=d.T.any)]), + withSizeLimit(sizeLimit): { spec+: { storage+: { emptyDir+: { sizeLimit: sizeLimit } } } }, + }, + '#ephemeral':: d.obj(help='"EphemeralVolumeSource to be used by the StatefulSet. This is a beta field in k8s 1.21 and GA in 1.15. For lower versions, starting with k8s 1.19, it requires enabling the GenericEphemeralVolume feature gate. More info: https://kubernetes.io/docs/concepts/storage/ephemeral-volumes/#generic-ephemeral-volumes"'), + ephemeral: { + '#volumeClaimTemplate':: d.obj(help='"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \\n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \\n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \\n Required, must not be nil."'), + volumeClaimTemplate: { + '#spec':: d.obj(help='"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } } } }, + }, + '#withMetadata':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadata(metadata): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { metadata: metadata } } } } }, + '#withMetadataMixin':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadataMixin(metadata): { spec+: { storage+: { ephemeral+: { volumeClaimTemplate+: { metadata+: metadata } } } } }, + }, + }, + '#volumeClaimTemplate':: d.obj(help='"Defines the PVC spec to be used by the Prometheus StatefulSets. The easiest way to use a volume that cannot be automatically provisioned is to use a label selector alongside manually created PersistentVolumes."'), + volumeClaimTemplate: { + '#metadata':: d.obj(help='"EmbeddedMetadata contains metadata relevant to an EmbeddedResource."'), + metadata: { + '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotations(annotations): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { annotations: annotations } } } } }, + '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]), + withAnnotationsMixin(annotations): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { annotations+: annotations } } } } }, + '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]), + withLabels(labels): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { labels: labels } } } } }, + '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]), + withLabelsMixin(labels): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { labels+: labels } } } } }, + '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { metadata+: { name: name } } } } }, + }, + '#spec':: d.obj(help='"Defines the desired characteristics of a volume requested by a pod author. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { spec+: { storage+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } } }, + }, + '#withApiVersion':: d.fn(help='"APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources"', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { spec+: { storage+: { volumeClaimTemplate+: { apiVersion: apiVersion } } } }, + '#withKind':: d.fn(help='"Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { spec+: { storage+: { volumeClaimTemplate+: { kind: kind } } } }, + }, + '#withDisableMountSubPath':: d.fn(help='"*Deprecated: subPath usage will be removed in a future release.*"', args=[d.arg(name='disableMountSubPath', type=d.T.boolean)]), + withDisableMountSubPath(disableMountSubPath): { spec+: { storage+: { disableMountSubPath: disableMountSubPath } } }, + }, + '#thanos':: d.obj(help='"Defines the configuration of the optional Thanos sidecar. \\n This section is experimental, it may change significantly without deprecation notice in any release."'), + thanos: { + '#additionalArgs':: d.obj(help='"AdditionalArgs allows setting additional arguments for the Thanos container. The arguments are passed as-is to the Thanos container which may cause issues if they are invalid or not supported the given Thanos version. In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged."'), + additionalArgs: { + '#withName':: d.fn(help='"Name of the argument, e.g. \\"scrape.discovery-reload-interval\\"."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withValue':: d.fn(help='"Argument value, e.g. 30s. Can be empty for name-only arguments (e.g. --storage.tsdb.no-lockfile)"', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#grpcServerTlsConfig':: d.obj(help='"Configures the TLS parameters for the gRPC server providing the StoreAPI. \\n Note: Currently only the `caFile`, `certFile`, and `keyFile` fields are supported."'), + grpcServerTlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { grpcServerTlsConfig+: { ca+: { secret+: { optional: optional } } } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { grpcServerTlsConfig+: { cert+: { secret+: { optional: optional } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { grpcServerTlsConfig+: { keySecret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { grpcServerTlsConfig+: { keySecret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { grpcServerTlsConfig+: { keySecret+: { optional: optional } } } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { spec+: { thanos+: { grpcServerTlsConfig+: { caFile: caFile } } } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { spec+: { thanos+: { grpcServerTlsConfig+: { certFile: certFile } } } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { spec+: { thanos+: { grpcServerTlsConfig+: { insecureSkipVerify: insecureSkipVerify } } } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { spec+: { thanos+: { grpcServerTlsConfig+: { keyFile: keyFile } } } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { spec+: { thanos+: { grpcServerTlsConfig+: { serverName: serverName } } } }, + }, + '#objectStorageConfig':: d.obj(help="\"Defines the Thanos sidecar's configuration to upload TSDB blocks to object storage. \\n More info: https://thanos.io/tip/thanos/storage.md/ \\n objectStorageConfigFile takes precedence over this field.\""), + objectStorageConfig: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { objectStorageConfig+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { objectStorageConfig+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { objectStorageConfig+: { optional: optional } } } }, + }, + '#resources':: d.obj(help='"Defines the resources requests and limits of the Thanos sidecar."'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { spec+: { thanos+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { spec+: { thanos+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { spec+: { thanos+: { resources+: { limits: limits } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { spec+: { thanos+: { resources+: { limits+: limits } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { spec+: { thanos+: { resources+: { requests: requests } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { spec+: { thanos+: { resources+: { requests+: requests } } } }, + }, + '#tracingConfig':: d.obj(help='"Defines the tracing configuration for the Thanos sidecar. \\n More info: https://thanos.io/tip/thanos/tracing.md/ \\n This is an experimental feature, it may change in any upcoming release in a breaking way. \\n tracingConfigFile takes precedence over this field."'), + tracingConfig: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { thanos+: { tracingConfig+: { key: key } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { thanos+: { tracingConfig+: { name: name } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { thanos+: { tracingConfig+: { optional: optional } } } }, + }, + '#volumeMounts':: d.obj(help="\"VolumeMounts allows configuration of additional VolumeMounts for Thanos. VolumeMounts specified will be appended to other VolumeMounts in the 'thanos-sidecar' container.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#withAdditionalArgs':: d.fn(help='"AdditionalArgs allows setting additional arguments for the Thanos container. The arguments are passed as-is to the Thanos container which may cause issues if they are invalid or not supported the given Thanos version. In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged."', args=[d.arg(name='additionalArgs', type=d.T.array)]), + withAdditionalArgs(additionalArgs): { spec+: { thanos+: { additionalArgs: if std.isArray(v=additionalArgs) then additionalArgs else [additionalArgs] } } }, + '#withAdditionalArgsMixin':: d.fn(help='"AdditionalArgs allows setting additional arguments for the Thanos container. The arguments are passed as-is to the Thanos container which may cause issues if they are invalid or not supported the given Thanos version. In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='additionalArgs', type=d.T.array)]), + withAdditionalArgsMixin(additionalArgs): { spec+: { thanos+: { additionalArgs+: if std.isArray(v=additionalArgs) then additionalArgs else [additionalArgs] } } }, + '#withBaseImage':: d.fn(help="\"*Deprecated: use 'image' instead.*\"", args=[d.arg(name='baseImage', type=d.T.string)]), + withBaseImage(baseImage): { spec+: { thanos+: { baseImage: baseImage } } }, + '#withBlockSize':: d.fn(help='"BlockDuration controls the size of TSDB blocks produced by Prometheus. The default value is 2h to match the upstream Prometheus defaults. \\n WARNING: Changing the block duration can impact the performance and efficiency of the entire Prometheus/Thanos stack due to how it interacts with memory and Thanos compactors. It is recommended to keep this value set to a multiple of 120 times your longest scrape or rule interval. For example, 30s * 120 = 1h."', args=[d.arg(name='blockSize', type=d.T.string)]), + withBlockSize(blockSize): { spec+: { thanos+: { blockSize: blockSize } } }, + '#withGetConfigInterval':: d.fn(help='"How often to retrieve the Prometheus configuration."', args=[d.arg(name='getConfigInterval', type=d.T.string)]), + withGetConfigInterval(getConfigInterval): { spec+: { thanos+: { getConfigInterval: getConfigInterval } } }, + '#withGetConfigTimeout':: d.fn(help='"Maximum time to wait when retrieving the Prometheus configuration."', args=[d.arg(name='getConfigTimeout', type=d.T.string)]), + withGetConfigTimeout(getConfigTimeout): { spec+: { thanos+: { getConfigTimeout: getConfigTimeout } } }, + '#withGrpcListenLocal':: d.fn(help="\"When true, the Thanos sidecar listens on the loopback interface instead of the Pod IP's address for the gRPC endpoints. \\n It has no effect if `listenLocal` is true.\"", args=[d.arg(name='grpcListenLocal', type=d.T.boolean)]), + withGrpcListenLocal(grpcListenLocal): { spec+: { thanos+: { grpcListenLocal: grpcListenLocal } } }, + '#withHttpListenLocal':: d.fn(help="\"When true, the Thanos sidecar listens on the loopback interface instead of the Pod IP's address for the HTTP endpoints. \\n It has no effect if `listenLocal` is true.\"", args=[d.arg(name='httpListenLocal', type=d.T.boolean)]), + withHttpListenLocal(httpListenLocal): { spec+: { thanos+: { httpListenLocal: httpListenLocal } } }, + '#withImage':: d.fn(help='"Container image name for Thanos. If specified, it takes precedence over the `spec.thanos.baseImage`, `spec.thanos.tag` and `spec.thanos.sha` fields. \\n Specifying `spec.thanos.version` is still necessary to ensure the Prometheus Operator knows which version of Thanos is being configured. \\n If neither `spec.thanos.image` nor `spec.thanos.baseImage` are defined, the operator will use the latest upstream version of Thanos available at the time when the operator was released."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { thanos+: { image: image } } }, + '#withListenLocal':: d.fn(help='"*Deprecated: use `grpcListenLocal` and `httpListenLocal` instead.*"', args=[d.arg(name='listenLocal', type=d.T.boolean)]), + withListenLocal(listenLocal): { spec+: { thanos+: { listenLocal: listenLocal } } }, + '#withLogFormat':: d.fn(help='"Log format for the Thanos sidecar."', args=[d.arg(name='logFormat', type=d.T.string)]), + withLogFormat(logFormat): { spec+: { thanos+: { logFormat: logFormat } } }, + '#withLogLevel':: d.fn(help='"Log level for the Thanos sidecar."', args=[d.arg(name='logLevel', type=d.T.string)]), + withLogLevel(logLevel): { spec+: { thanos+: { logLevel: logLevel } } }, + '#withMinTime':: d.fn(help="\"Defines the start of time range limit served by the Thanos sidecar's StoreAPI. The field's value should be a constant time in RFC3339 format or a time duration relative to current time, such as -1d or 2h45m. Valid duration units are ms, s, m, h, d, w, y.\"", args=[d.arg(name='minTime', type=d.T.string)]), + withMinTime(minTime): { spec+: { thanos+: { minTime: minTime } } }, + '#withObjectStorageConfigFile':: d.fn(help="\"Defines the Thanos sidecar's configuration file to upload TSDB blocks to object storage. \\n More info: https://thanos.io/tip/thanos/storage.md/ \\n This field takes precedence over objectStorageConfig.\"", args=[d.arg(name='objectStorageConfigFile', type=d.T.string)]), + withObjectStorageConfigFile(objectStorageConfigFile): { spec+: { thanos+: { objectStorageConfigFile: objectStorageConfigFile } } }, + '#withReadyTimeout':: d.fn(help='"ReadyTimeout is the maximum time that the Thanos sidecar will wait for Prometheus to start."', args=[d.arg(name='readyTimeout', type=d.T.string)]), + withReadyTimeout(readyTimeout): { spec+: { thanos+: { readyTimeout: readyTimeout } } }, + '#withSha':: d.fn(help="\"*Deprecated: use 'image' instead. The image digest can be specified as part of the image name.*\"", args=[d.arg(name='sha', type=d.T.string)]), + withSha(sha): { spec+: { thanos+: { sha: sha } } }, + '#withTag':: d.fn(help="\"*Deprecated: use 'image' instead. The image's tag can be specified as part of the image name.*\"", args=[d.arg(name='tag', type=d.T.string)]), + withTag(tag): { spec+: { thanos+: { tag: tag } } }, + '#withTracingConfigFile':: d.fn(help='"Defines the tracing configuration file for the Thanos sidecar. \\n More info: https://thanos.io/tip/thanos/tracing.md/ \\n This is an experimental feature, it may change in any upcoming release in a breaking way. \\n This field takes precedence over tracingConfig."', args=[d.arg(name='tracingConfigFile', type=d.T.string)]), + withTracingConfigFile(tracingConfigFile): { spec+: { thanos+: { tracingConfigFile: tracingConfigFile } } }, + '#withVersion':: d.fn(help='"Version of Thanos being deployed. The operator uses this information to generate the Prometheus StatefulSet + configuration files. \\n If not specified, the operator assumes the latest upstream release of Thanos available at the time when the version of the operator was released."', args=[d.arg(name='version', type=d.T.string)]), + withVersion(version): { spec+: { thanos+: { version: version } } }, + '#withVolumeMounts':: d.fn(help="\"VolumeMounts allows configuration of additional VolumeMounts for Thanos. VolumeMounts specified will be appended to other VolumeMounts in the 'thanos-sidecar' container.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMounts(volumeMounts): { spec+: { thanos+: { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } } }, + '#withVolumeMountsMixin':: d.fn(help="\"VolumeMounts allows configuration of additional VolumeMounts for Thanos. VolumeMounts specified will be appended to other VolumeMounts in the 'thanos-sidecar' container.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]), + withVolumeMountsMixin(volumeMounts): { spec+: { thanos+: { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] } } }, + }, + '#tolerations':: d.obj(help="\"Defines the Pods' tolerations if specified.\""), + tolerations: { + '#withEffect':: d.fn(help='"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."', args=[d.arg(name='effect', type=d.T.string)]), + withEffect(effect): { effect: effect }, + '#withKey':: d.fn(help='"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withTolerationSeconds':: d.fn(help='"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."', args=[d.arg(name='tolerationSeconds', type=d.T.integer)]), + withTolerationSeconds(tolerationSeconds): { tolerationSeconds: tolerationSeconds }, + '#withValue':: d.fn(help='"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."', args=[d.arg(name='value', type=d.T.string)]), + withValue(value): { value: value }, + }, + '#topologySpreadConstraints':: d.obj(help="\"Defines the pod's topology spread constraints if specified.\""), + topologySpreadConstraints: { + '#labelSelector':: d.obj(help='"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."'), + labelSelector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } }, + }, + '#withMatchLabelKeys':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='matchLabelKeys', type=d.T.array)]), + withMatchLabelKeys(matchLabelKeys): { matchLabelKeys: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] }, + '#withMatchLabelKeysMixin':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchLabelKeys', type=d.T.array)]), + withMatchLabelKeysMixin(matchLabelKeys): { matchLabelKeys+: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] }, + '#withMaxSkew':: d.fn(help="\"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed.\"", args=[d.arg(name='maxSkew', type=d.T.integer)]), + withMaxSkew(maxSkew): { maxSkew: maxSkew }, + '#withMinDomains':: d.fn(help="\"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \\\"global minimum\\\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \\n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \\\"global minimum\\\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \\n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='minDomains', type=d.T.integer)]), + withMinDomains(minDomains): { minDomains: minDomains }, + '#withNodeAffinityPolicy':: d.fn(help="\"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \\n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag.\"", args=[d.arg(name='nodeAffinityPolicy', type=d.T.string)]), + withNodeAffinityPolicy(nodeAffinityPolicy): { nodeAffinityPolicy: nodeAffinityPolicy }, + '#withNodeTaintsPolicy':: d.fn(help='"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \\n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."', args=[d.arg(name='nodeTaintsPolicy', type=d.T.string)]), + withNodeTaintsPolicy(nodeTaintsPolicy): { nodeTaintsPolicy: nodeTaintsPolicy }, + '#withTopologyKey':: d.fn(help="\"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each \u003ckey, value\u003e as a \\\"bucket\\\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \\\"kubernetes.io/hostname\\\", each Node is a domain of that topology. And, if TopologyKey is \\\"topology.kubernetes.io/zone\\\", each zone is a domain of that topology. It's a required field.\"", args=[d.arg(name='topologyKey', type=d.T.string)]), + withTopologyKey(topologyKey): { topologyKey: topologyKey }, + '#withWhenUnsatisfiable':: d.fn(help="\"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \\\"Unsatisfiable\\\" for an incoming pod if and only if every possible node assignment for that pod would violate \\\"MaxSkew\\\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field.\"", args=[d.arg(name='whenUnsatisfiable', type=d.T.string)]), + withWhenUnsatisfiable(whenUnsatisfiable): { whenUnsatisfiable: whenUnsatisfiable }, + }, + '#tracingConfig':: d.obj(help='"EXPERIMENTAL: TracingConfig configures tracing in Prometheus. This is an experimental feature, it may change in any upcoming release in a breaking way."'), + tracingConfig: { + '#tlsConfig':: d.obj(help='"TLS Config to use when sending traces."'), + tlsConfig: { + '#ca':: d.obj(help='"Certificate authority used when verifying server certificates."'), + ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tracingConfig+: { tlsConfig+: { ca+: { secret+: { optional: optional } } } } } }, + }, + }, + '#cert':: d.obj(help='"Client certificate to present when doing client-authentication."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tracingConfig+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the client key file for the targets."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { tracingConfig+: { tlsConfig+: { keySecret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { tracingConfig+: { tlsConfig+: { keySecret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { tracingConfig+: { tlsConfig+: { keySecret+: { optional: optional } } } } }, + }, + '#withCaFile':: d.fn(help='"Path to the CA cert in the Prometheus container to use for the targets."', args=[d.arg(name='caFile', type=d.T.string)]), + withCaFile(caFile): { spec+: { tracingConfig+: { tlsConfig+: { caFile: caFile } } } }, + '#withCertFile':: d.fn(help='"Path to the client cert file in the Prometheus container for the targets."', args=[d.arg(name='certFile', type=d.T.string)]), + withCertFile(certFile): { spec+: { tracingConfig+: { tlsConfig+: { certFile: certFile } } } }, + '#withInsecureSkipVerify':: d.fn(help='"Disable target certificate validation."', args=[d.arg(name='insecureSkipVerify', type=d.T.boolean)]), + withInsecureSkipVerify(insecureSkipVerify): { spec+: { tracingConfig+: { tlsConfig+: { insecureSkipVerify: insecureSkipVerify } } } }, + '#withKeyFile':: d.fn(help='"Path to the client key file in the Prometheus container for the targets."', args=[d.arg(name='keyFile', type=d.T.string)]), + withKeyFile(keyFile): { spec+: { tracingConfig+: { tlsConfig+: { keyFile: keyFile } } } }, + '#withServerName':: d.fn(help='"Used to verify the hostname for the targets."', args=[d.arg(name='serverName', type=d.T.string)]), + withServerName(serverName): { spec+: { tracingConfig+: { tlsConfig+: { serverName: serverName } } } }, + }, + '#withClientType':: d.fn(help='"Client used to export the traces. Supported values are `http` or `grpc`."', args=[d.arg(name='clientType', type=d.T.string)]), + withClientType(clientType): { spec+: { tracingConfig+: { clientType: clientType } } }, + '#withCompression':: d.fn(help='"Compression key for supported compression types. The only supported value is `gzip`."', args=[d.arg(name='compression', type=d.T.string)]), + withCompression(compression): { spec+: { tracingConfig+: { compression: compression } } }, + '#withEndpoint':: d.fn(help='"Endpoint to send the traces to. Should be provided in format :."', args=[d.arg(name='endpoint', type=d.T.string)]), + withEndpoint(endpoint): { spec+: { tracingConfig+: { endpoint: endpoint } } }, + '#withHeaders':: d.fn(help='"Key-value pairs to be used as headers associated with gRPC or HTTP requests."', args=[d.arg(name='headers', type=d.T.object)]), + withHeaders(headers): { spec+: { tracingConfig+: { headers: headers } } }, + '#withHeadersMixin':: d.fn(help='"Key-value pairs to be used as headers associated with gRPC or HTTP requests."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='headers', type=d.T.object)]), + withHeadersMixin(headers): { spec+: { tracingConfig+: { headers+: headers } } }, + '#withInsecure':: d.fn(help='"If disabled, the client will use a secure connection."', args=[d.arg(name='insecure', type=d.T.boolean)]), + withInsecure(insecure): { spec+: { tracingConfig+: { insecure: insecure } } }, + '#withSamplingFraction':: d.fn(help='"Sets the probability a given trace will be sampled. Must be a float from 0 through 1."', args=[d.arg(name='samplingFraction', type=d.T.any)]), + withSamplingFraction(samplingFraction): { spec+: { tracingConfig+: { samplingFraction: samplingFraction } } }, + '#withTimeout':: d.fn(help='"Maximum time the exporter will wait for each batch export."', args=[d.arg(name='timeout', type=d.T.string)]), + withTimeout(timeout): { spec+: { tracingConfig+: { timeout: timeout } } }, + }, + '#tsdb':: d.obj(help='"Defines the runtime reloadable configuration of the timeseries database (TSDB)."'), + tsdb: { + '#withOutOfOrderTimeWindow':: d.fn(help='"Configures how old an out-of-order/out-of-bounds sample can be with respect to the TSDB max time. \\n An out-of-order/out-of-bounds sample is ingested into the TSDB as long as the timestamp of the sample is >= (TSDB.MaxTime - outOfOrderTimeWindow). \\n Out of order ingestion is an experimental feature. \\n It requires Prometheus >= v2.39.0."', args=[d.arg(name='outOfOrderTimeWindow', type=d.T.string)]), + withOutOfOrderTimeWindow(outOfOrderTimeWindow): { spec+: { tsdb+: { outOfOrderTimeWindow: outOfOrderTimeWindow } } }, + }, + '#volumeMounts':: d.obj(help="\"VolumeMounts allows the configuration of additional VolumeMounts. \\n VolumeMounts will be appended to other VolumeMounts in the 'prometheus' container, that are generated as a result of StorageSpec objects.\""), + volumeMounts: { + '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]), + withMountPath(mountPath): { mountPath: mountPath }, + '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]), + withMountPropagation(mountPropagation): { mountPropagation: mountPropagation }, + '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { readOnly: readOnly }, + '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]), + withSubPath(subPath): { subPath: subPath }, + '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]), + withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr }, + }, + '#volumes':: d.obj(help='"Volumes allows the configuration of additional volumes on the output StatefulSet definition. Volumes specified will be appended to other volumes that are generated as a result of StorageSpec objects."'), + volumes: { + '#awsElasticBlockStore':: d.obj(help="\"awsElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore\""), + awsElasticBlockStore: { + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { awsElasticBlockStore+: { fsType: fsType } }, + '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty)."', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { awsElasticBlockStore+: { partition: partition } }, + '#withReadOnly':: d.fn(help='"readOnly value true will force the readOnly setting in VolumeMounts. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { awsElasticBlockStore+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID is unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { awsElasticBlockStore+: { volumeID: volumeID } }, + }, + '#azureDisk':: d.obj(help='"azureDisk represents an Azure Data Disk mount on the host and bind mount to the pod."'), + azureDisk: { + '#withCachingMode':: d.fn(help='"cachingMode is the Host Caching mode: None, Read Only, Read Write."', args=[d.arg(name='cachingMode', type=d.T.string)]), + withCachingMode(cachingMode): { azureDisk+: { cachingMode: cachingMode } }, + '#withDiskName':: d.fn(help='"diskName is the Name of the data disk in the blob storage"', args=[d.arg(name='diskName', type=d.T.string)]), + withDiskName(diskName): { azureDisk+: { diskName: diskName } }, + '#withDiskURI':: d.fn(help='"diskURI is the URI of data disk in the blob storage"', args=[d.arg(name='diskURI', type=d.T.string)]), + withDiskURI(diskURI): { azureDisk+: { diskURI: diskURI } }, + '#withFsType':: d.fn(help='"fsType is Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { azureDisk+: { fsType: fsType } }, + '#withKind':: d.fn(help='"kind expected values are Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { azureDisk+: { kind: kind } }, + '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureDisk+: { readOnly: readOnly } }, + }, + '#azureFile':: d.obj(help='"azureFile represents an Azure File Service mount on the host and bind mount to the pod."'), + azureFile: { + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { azureFile+: { readOnly: readOnly } }, + '#withSecretName':: d.fn(help='"secretName is the name of secret that contains Azure Storage Account Name and Key"', args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { azureFile+: { secretName: secretName } }, + '#withShareName':: d.fn(help='"shareName is the azure share Name"', args=[d.arg(name='shareName', type=d.T.string)]), + withShareName(shareName): { azureFile+: { shareName: shareName } }, + }, + '#cephfs':: d.obj(help="\"cephFS represents a Ceph FS mount on the host that shares a pod's lifetime\""), + cephfs: { + '#secretRef':: d.obj(help='"secretRef is Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cephfs+: { secretRef+: { name: name } } }, + }, + '#withMonitors':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPath':: d.fn(help='"path is Optional: Used as the mounted root, rather than the full Ceph tree, default is /"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { cephfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cephfs+: { readOnly: readOnly } }, + '#withSecretFile':: d.fn(help='"secretFile is Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='secretFile', type=d.T.string)]), + withSecretFile(secretFile): { cephfs+: { secretFile: secretFile } }, + '#withUser':: d.fn(help='"user is optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { cephfs+: { user: user } }, + }, + '#cinder':: d.obj(help='"cinder represents a cinder volume attached and mounted on kubelets host machine. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"'), + cinder: { + '#secretRef':: d.obj(help='"secretRef is optional: points to a secret object containing parameters used to connect to OpenStack."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { cinder+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { cinder+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { cinder+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { cinder+: { volumeID: volumeID } }, + }, + '#configMap':: d.obj(help='"configMap represents a configMap that should populate this volume"'), + configMap: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode is optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { configMap+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#csi':: d.obj(help='"csi (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature)."'), + csi: { + '#nodePublishSecretRef':: d.obj(help='"nodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed."'), + nodePublishSecretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { csi+: { nodePublishSecretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='"driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster."', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { csi+: { driver: driver } }, + '#withFsType':: d.fn(help='"fsType to mount. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { csi+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly specifies a read-only configuration for the volume. Defaults to false (read/write)."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { csi+: { readOnly: readOnly } }, + '#withVolumeAttributes':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributes(volumeAttributes): { csi+: { volumeAttributes: volumeAttributes } }, + '#withVolumeAttributesMixin':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeAttributes', type=d.T.object)]), + withVolumeAttributesMixin(volumeAttributes): { csi+: { volumeAttributes+: volumeAttributes } }, + }, + '#downwardAPI':: d.obj(help='"downwardAPI represents downward API about the pod that should populate this volume"'), + downwardAPI: { + '#items':: d.obj(help='"Items is a list of downward API volume file"'), + items: { + '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { downwardAPI+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help='"Items is a list of downward API volume file"', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='"Items is a list of downward API volume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#emptyDir':: d.obj(help="\"emptyDir represents a temporary directory that shares a pod's lifetime. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\""), + emptyDir: { + '#withMedium':: d.fn(help="\"medium represents what type of storage medium should back this directory. The default is \\\"\\\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\"", args=[d.arg(name='medium', type=d.T.string)]), + withMedium(medium): { emptyDir+: { medium: medium } }, + '#withSizeLimit':: d.fn(help='"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"', args=[d.arg(name='sizeLimit', type=d.T.any)]), + withSizeLimit(sizeLimit): { emptyDir+: { sizeLimit: sizeLimit } }, + }, + '#ephemeral':: d.obj(help="\"ephemeral represents a volume that is handled by a cluster storage driver. The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, and deleted when the pod is removed. \\n Use this if: a) the volume is only needed while the pod runs, b) features of normal volumes like restoring from snapshot or capacity tracking are needed, c) the storage driver is specified through a storage class, and d) the storage driver supports dynamic volume provisioning through a PersistentVolumeClaim (see EphemeralVolumeSource for more information on the connection between this volume type and PersistentVolumeClaim). \\n Use PersistentVolumeClaim or one of the vendor-specific APIs for volumes that persist for longer than the lifecycle of an individual pod. \\n Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to be used that way - see the documentation of the driver for more information. \\n A pod can use both types of ephemeral volumes and persistent volumes at the same time.\""), + ephemeral: { + '#volumeClaimTemplate':: d.obj(help='"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \\n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \\n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \\n Required, must not be nil."'), + volumeClaimTemplate: { + '#spec':: d.obj(help='"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."'), + spec: { + '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'), + dataSource: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } }, + }, + '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""), + dataSourceRef: { + '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]), + withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } }, + '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]), + withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } }, + '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } }, + '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]), + withNamespace(namespace): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } }, + }, + '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'), + resources: { + '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'), + claims: { + '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]), + withClaims(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } }, + '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]), + withClaimsMixin(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } }, + '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]), + withLimits(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } }, + '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]), + withLimitsMixin(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } }, + '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]), + withRequests(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } }, + '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]), + withRequestsMixin(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } }, + }, + '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'), + selector: { + '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'), + matchExpressions: { + '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]), + withOperator(operator): { operator: operator }, + '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]), + withValues(values): { values: if std.isArray(v=values) then values else [values] }, + '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]), + withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] }, + }, + '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressions(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]), + withMatchExpressionsMixin(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } }, + '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabels(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } }, + '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]), + withMatchLabelsMixin(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } }, + }, + '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModes(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]), + withAccessModesMixin(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } }, + '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]), + withStorageClassName(storageClassName): { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } }, + '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]), + withVolumeMode(volumeMode): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } }, + '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } }, + }, + '#withMetadata':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadata(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata: metadata } } }, + '#withMetadataMixin':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metadata', type=d.T.object)]), + withMetadataMixin(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata+: metadata } } }, + }, + }, + '#fc':: d.obj(help="\"fc represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod.\""), + fc: { + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { fc+: { fsType: fsType } }, + '#withLun':: d.fn(help='"lun is Optional: FC target lun number"', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { fc+: { lun: lun } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { fc+: { readOnly: readOnly } }, + '#withTargetWWNs':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNs(targetWWNs): { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withTargetWWNsMixin':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]), + withTargetWWNsMixin(targetWWNs): { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } }, + '#withWwids':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."', args=[d.arg(name='wwids', type=d.T.array)]), + withWwids(wwids): { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } }, + '#withWwidsMixin':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]), + withWwidsMixin(wwids): { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } }, + }, + '#flexVolume':: d.obj(help='"flexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin."'), + flexVolume: { + '#secretRef':: d.obj(help='"secretRef is Optional: secretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { flexVolume+: { secretRef+: { name: name } } }, + }, + '#withDriver':: d.fn(help='"driver is the name of the driver to use for this volume."', args=[d.arg(name='driver', type=d.T.string)]), + withDriver(driver): { flexVolume+: { driver: driver } }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". The default filesystem depends on FlexVolume script."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { flexVolume+: { fsType: fsType } }, + '#withOptions':: d.fn(help='"options is Optional: this field holds extra command options if any."', args=[d.arg(name='options', type=d.T.object)]), + withOptions(options): { flexVolume+: { options: options } }, + '#withOptionsMixin':: d.fn(help='"options is Optional: this field holds extra command options if any."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]), + withOptionsMixin(options): { flexVolume+: { options+: options } }, + '#withReadOnly':: d.fn(help='"readOnly is Optional: defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { flexVolume+: { readOnly: readOnly } }, + }, + '#flocker':: d.obj(help="\"flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running\""), + flocker: { + '#withDatasetName':: d.fn(help='"datasetName is Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated"', args=[d.arg(name='datasetName', type=d.T.string)]), + withDatasetName(datasetName): { flocker+: { datasetName: datasetName } }, + '#withDatasetUUID':: d.fn(help='"datasetUUID is the UUID of the dataset. This is unique identifier of a Flocker dataset"', args=[d.arg(name='datasetUUID', type=d.T.string)]), + withDatasetUUID(datasetUUID): { flocker+: { datasetUUID: datasetUUID } }, + }, + '#gcePersistentDisk':: d.obj(help="\"gcePersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk\""), + gcePersistentDisk: { + '#withFsType':: d.fn(help='"fsType is filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { gcePersistentDisk+: { fsType: fsType } }, + '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='partition', type=d.T.integer)]), + withPartition(partition): { gcePersistentDisk+: { partition: partition } }, + '#withPdName':: d.fn(help='"pdName is unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='pdName', type=d.T.string)]), + withPdName(pdName): { gcePersistentDisk+: { pdName: pdName } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { gcePersistentDisk+: { readOnly: readOnly } }, + }, + '#gitRepo':: d.obj(help="\"gitRepo represents a git repository at a particular revision. DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container.\""), + gitRepo: { + '#withDirectory':: d.fn(help="\"directory is the target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name.\"", args=[d.arg(name='directory', type=d.T.string)]), + withDirectory(directory): { gitRepo+: { directory: directory } }, + '#withRepository':: d.fn(help='"repository is the URL"', args=[d.arg(name='repository', type=d.T.string)]), + withRepository(repository): { gitRepo+: { repository: repository } }, + '#withRevision':: d.fn(help='"revision is the commit hash for the specified revision."', args=[d.arg(name='revision', type=d.T.string)]), + withRevision(revision): { gitRepo+: { revision: revision } }, + }, + '#glusterfs':: d.obj(help="\"glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/glusterfs/README.md\""), + glusterfs: { + '#withEndpoints':: d.fn(help='"endpoints is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='endpoints', type=d.T.string)]), + withEndpoints(endpoints): { glusterfs+: { endpoints: endpoints } }, + '#withPath':: d.fn(help='"path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { glusterfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { glusterfs+: { readOnly: readOnly } }, + }, + '#hostPath':: d.obj(help='"hostPath represents a pre-existing file or directory on the host machine that is directly exposed to the container. This is generally used for system agents or other privileged things that are allowed to see the host machine. Most containers will NOT need this. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath --- TODO(jonesdl) We need to restrict who can use host directory mounts and who can/can not mount host directories as read/write."'), + hostPath: { + '#withPath':: d.fn(help='"path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { hostPath+: { path: path } }, + '#withType':: d.fn(help='"type for HostPath Volume Defaults to \\"\\" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='type', type=d.T.string)]), + withType(type): { hostPath+: { type: type } }, + }, + '#iscsi':: d.obj(help="\"iscsi represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://examples.k8s.io/volumes/iscsi/README.md\""), + iscsi: { + '#secretRef':: d.obj(help='"secretRef is the CHAP Secret for iSCSI target and initiator authentication"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { iscsi+: { secretRef+: { name: name } } }, + }, + '#withChapAuthDiscovery':: d.fn(help='"chapAuthDiscovery defines whether support iSCSI Discovery CHAP authentication"', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]), + withChapAuthDiscovery(chapAuthDiscovery): { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } }, + '#withChapAuthSession':: d.fn(help='"chapAuthSession defines whether support iSCSI Session CHAP authentication"', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]), + withChapAuthSession(chapAuthSession): { iscsi+: { chapAuthSession: chapAuthSession } }, + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { iscsi+: { fsType: fsType } }, + '#withInitiatorName':: d.fn(help='"initiatorName is the custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection."', args=[d.arg(name='initiatorName', type=d.T.string)]), + withInitiatorName(initiatorName): { iscsi+: { initiatorName: initiatorName } }, + '#withIqn':: d.fn(help='"iqn is the target iSCSI Qualified Name."', args=[d.arg(name='iqn', type=d.T.string)]), + withIqn(iqn): { iscsi+: { iqn: iqn } }, + '#withIscsiInterface':: d.fn(help="\"iscsiInterface is the interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).\"", args=[d.arg(name='iscsiInterface', type=d.T.string)]), + withIscsiInterface(iscsiInterface): { iscsi+: { iscsiInterface: iscsiInterface } }, + '#withLun':: d.fn(help='"lun represents iSCSI Target Lun number."', args=[d.arg(name='lun', type=d.T.integer)]), + withLun(lun): { iscsi+: { lun: lun } }, + '#withPortals':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='portals', type=d.T.array)]), + withPortals(portals): { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } }, + '#withPortalsMixin':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]), + withPortalsMixin(portals): { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { iscsi+: { readOnly: readOnly } }, + '#withTargetPortal':: d.fn(help='"targetPortal is iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='targetPortal', type=d.T.string)]), + withTargetPortal(targetPortal): { iscsi+: { targetPortal: targetPortal } }, + }, + '#nfs':: d.obj(help="\"nfs represents an NFS mount on the host that shares a pod's lifetime More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs\""), + nfs: { + '#withPath':: d.fn(help='"path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { nfs+: { path: path } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { nfs+: { readOnly: readOnly } }, + '#withServer':: d.fn(help='"server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='server', type=d.T.string)]), + withServer(server): { nfs+: { server: server } }, + }, + '#persistentVolumeClaim':: d.obj(help='"persistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"'), + persistentVolumeClaim: { + '#withClaimName':: d.fn(help='"claimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"', args=[d.arg(name='claimName', type=d.T.string)]), + withClaimName(claimName): { persistentVolumeClaim+: { claimName: claimName } }, + '#withReadOnly':: d.fn(help='"readOnly Will force the ReadOnly setting in VolumeMounts. Default false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { persistentVolumeClaim+: { readOnly: readOnly } }, + }, + '#photonPersistentDisk':: d.obj(help='"photonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine"'), + photonPersistentDisk: { + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { photonPersistentDisk+: { fsType: fsType } }, + '#withPdID':: d.fn(help='"pdID is the ID that identifies Photon Controller persistent disk"', args=[d.arg(name='pdID', type=d.T.string)]), + withPdID(pdID): { photonPersistentDisk+: { pdID: pdID } }, + }, + '#portworxVolume':: d.obj(help='"portworxVolume represents a portworx volume attached and mounted on kubelets host machine"'), + portworxVolume: { + '#withFsType':: d.fn(help='"fSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { portworxVolume+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { portworxVolume+: { readOnly: readOnly } }, + '#withVolumeID':: d.fn(help='"volumeID uniquely identifies a Portworx volume"', args=[d.arg(name='volumeID', type=d.T.string)]), + withVolumeID(volumeID): { portworxVolume+: { volumeID: volumeID } }, + }, + '#projected':: d.obj(help='"projected items for all in one resources secrets, configmaps, and downward API"'), + projected: { + '#sources':: d.obj(help='"sources is the list of volume projections"'), + sources: { + '#configMap':: d.obj(help='"configMap information about the configMap data to project"'), + configMap: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { configMap+: { name: name } }, + '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { configMap+: { optional: optional } }, + }, + '#downwardAPI':: d.obj(help='"downwardAPI information about the downwardAPI data to project"'), + downwardAPI: { + '#items':: d.obj(help='"Items is a list of DownwardAPIVolume file"'), + items: { + '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'), + fieldRef: { + '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]), + withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } }, + '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]), + withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } }, + }, + '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'), + resourceFieldRef: { + '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]), + withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } }, + '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]), + withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } }, + '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]), + withResource(resource): { resourceFieldRef+: { resource: resource } }, + }, + '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help='"Items is a list of DownwardAPIVolume file"', args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help='"Items is a list of DownwardAPIVolume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } }, + }, + '#secret':: d.obj(help='"secret information about the secret data to project"'), + secret: { + '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { secret+: { name: name } }, + '#withOptional':: d.fn(help='"optional field specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + }, + '#serviceAccountToken':: d.obj(help='"serviceAccountToken is information about the serviceAccountToken data to project"'), + serviceAccountToken: { + '#withAudience':: d.fn(help='"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."', args=[d.arg(name='audience', type=d.T.string)]), + withAudience(audience): { serviceAccountToken+: { audience: audience } }, + '#withExpirationSeconds':: d.fn(help='"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."', args=[d.arg(name='expirationSeconds', type=d.T.integer)]), + withExpirationSeconds(expirationSeconds): { serviceAccountToken+: { expirationSeconds: expirationSeconds } }, + '#withPath':: d.fn(help='"path is the path relative to the mount point of the file to project the token into."', args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { serviceAccountToken+: { path: path } }, + }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { projected+: { defaultMode: defaultMode } }, + '#withSources':: d.fn(help='"sources is the list of volume projections"', args=[d.arg(name='sources', type=d.T.array)]), + withSources(sources): { projected+: { sources: if std.isArray(v=sources) then sources else [sources] } }, + '#withSourcesMixin':: d.fn(help='"sources is the list of volume projections"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]), + withSourcesMixin(sources): { projected+: { sources+: if std.isArray(v=sources) then sources else [sources] } }, + }, + '#quobyte':: d.obj(help="\"quobyte represents a Quobyte mount on the host that shares a pod's lifetime\""), + quobyte: { + '#withGroup':: d.fn(help='"group to map volume access to Default is no group"', args=[d.arg(name='group', type=d.T.string)]), + withGroup(group): { quobyte+: { group: group } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { quobyte+: { readOnly: readOnly } }, + '#withRegistry':: d.fn(help='"registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes"', args=[d.arg(name='registry', type=d.T.string)]), + withRegistry(registry): { quobyte+: { registry: registry } }, + '#withTenant':: d.fn(help='"tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin"', args=[d.arg(name='tenant', type=d.T.string)]), + withTenant(tenant): { quobyte+: { tenant: tenant } }, + '#withUser':: d.fn(help='"user to map volume access to Defaults to serivceaccount user"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { quobyte+: { user: user } }, + '#withVolume':: d.fn(help='"volume is a string that references an already created Quobyte volume by name."', args=[d.arg(name='volume', type=d.T.string)]), + withVolume(volume): { quobyte+: { volume: volume } }, + }, + '#rbd':: d.obj(help="\"rbd represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/rbd/README.md\""), + rbd: { + '#secretRef':: d.obj(help='"secretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { rbd+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { rbd+: { fsType: fsType } }, + '#withImage':: d.fn(help='"image is the rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { rbd+: { image: image } }, + '#withKeyring':: d.fn(help='"keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='keyring', type=d.T.string)]), + withKeyring(keyring): { rbd+: { keyring: keyring } }, + '#withMonitors':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitors(monitors): { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withMonitorsMixin':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]), + withMonitorsMixin(monitors): { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } }, + '#withPool':: d.fn(help='"pool is the rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='pool', type=d.T.string)]), + withPool(pool): { rbd+: { pool: pool } }, + '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { rbd+: { readOnly: readOnly } }, + '#withUser':: d.fn(help='"user is the rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]), + withUser(user): { rbd+: { user: user } }, + }, + '#scaleIO':: d.obj(help='"scaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes."'), + scaleIO: { + '#secretRef':: d.obj(help='"secretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { scaleIO+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Default is \\"xfs\\"."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { scaleIO+: { fsType: fsType } }, + '#withGateway':: d.fn(help='"gateway is the host address of the ScaleIO API Gateway."', args=[d.arg(name='gateway', type=d.T.string)]), + withGateway(gateway): { scaleIO+: { gateway: gateway } }, + '#withProtectionDomain':: d.fn(help='"protectionDomain is the name of the ScaleIO Protection Domain for the configured storage."', args=[d.arg(name='protectionDomain', type=d.T.string)]), + withProtectionDomain(protectionDomain): { scaleIO+: { protectionDomain: protectionDomain } }, + '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { scaleIO+: { readOnly: readOnly } }, + '#withSslEnabled':: d.fn(help='"sslEnabled Flag enable/disable SSL communication with Gateway, default false"', args=[d.arg(name='sslEnabled', type=d.T.boolean)]), + withSslEnabled(sslEnabled): { scaleIO+: { sslEnabled: sslEnabled } }, + '#withStorageMode':: d.fn(help='"storageMode indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned."', args=[d.arg(name='storageMode', type=d.T.string)]), + withStorageMode(storageMode): { scaleIO+: { storageMode: storageMode } }, + '#withStoragePool':: d.fn(help='"storagePool is the ScaleIO Storage Pool associated with the protection domain."', args=[d.arg(name='storagePool', type=d.T.string)]), + withStoragePool(storagePool): { scaleIO+: { storagePool: storagePool } }, + '#withSystem':: d.fn(help='"system is the name of the storage system as configured in ScaleIO."', args=[d.arg(name='system', type=d.T.string)]), + withSystem(system): { scaleIO+: { system: system } }, + '#withVolumeName':: d.fn(help='"volumeName is the name of a volume already created in the ScaleIO system that is associated with this volume source."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { scaleIO+: { volumeName: volumeName } }, + }, + '#secret':: d.obj(help='"secret represents a secret that should populate this volume. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret"'), + secret: { + '#items':: d.obj(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""), + items: { + '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { key: key }, + '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]), + withMode(mode): { mode: mode }, + '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]), + withPath(path): { path: path }, + }, + '#withDefaultMode':: d.fn(help='"defaultMode is Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]), + withDefaultMode(defaultMode): { secret+: { defaultMode: defaultMode } }, + '#withItems':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]), + withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } }, + '#withItemsMixin':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]), + withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } }, + '#withOptional':: d.fn(help='"optional field specify whether the Secret or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { secret+: { optional: optional } }, + '#withSecretName':: d.fn(help="\"secretName is the name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret\"", args=[d.arg(name='secretName', type=d.T.string)]), + withSecretName(secretName): { secret+: { secretName: secretName } }, + }, + '#storageos':: d.obj(help='"storageOS represents a StorageOS volume attached and mounted on Kubernetes nodes."'), + storageos: { + '#secretRef':: d.obj(help='"secretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted."'), + secretRef: { + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { storageos+: { secretRef+: { name: name } } }, + }, + '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { storageos+: { fsType: fsType } }, + '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]), + withReadOnly(readOnly): { storageos+: { readOnly: readOnly } }, + '#withVolumeName':: d.fn(help='"volumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace."', args=[d.arg(name='volumeName', type=d.T.string)]), + withVolumeName(volumeName): { storageos+: { volumeName: volumeName } }, + '#withVolumeNamespace':: d.fn(help="\"volumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to \\\"default\\\" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.\"", args=[d.arg(name='volumeNamespace', type=d.T.string)]), + withVolumeNamespace(volumeNamespace): { storageos+: { volumeNamespace: volumeNamespace } }, + }, + '#vsphereVolume':: d.obj(help='"vsphereVolume represents a vSphere volume attached and mounted on kubelets host machine"'), + vsphereVolume: { + '#withFsType':: d.fn(help='"fsType is filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]), + withFsType(fsType): { vsphereVolume+: { fsType: fsType } }, + '#withStoragePolicyID':: d.fn(help='"storagePolicyID is the storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName."', args=[d.arg(name='storagePolicyID', type=d.T.string)]), + withStoragePolicyID(storagePolicyID): { vsphereVolume+: { storagePolicyID: storagePolicyID } }, + '#withStoragePolicyName':: d.fn(help='"storagePolicyName is the storage Policy Based Management (SPBM) profile name."', args=[d.arg(name='storagePolicyName', type=d.T.string)]), + withStoragePolicyName(storagePolicyName): { vsphereVolume+: { storagePolicyName: storagePolicyName } }, + '#withVolumePath':: d.fn(help='"volumePath is the path that identifies vSphere volume vmdk"', args=[d.arg(name='volumePath', type=d.T.string)]), + withVolumePath(volumePath): { vsphereVolume+: { volumePath: volumePath } }, + }, + '#withName':: d.fn(help='"name of the volume. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { name: name }, + }, + '#web':: d.obj(help='"Defines the configuration of the Prometheus web server."'), + web: { + '#httpConfig':: d.obj(help='"Defines HTTP parameters for web server."'), + httpConfig: { + '#headers':: d.obj(help='"List of headers that can be added to HTTP responses."'), + headers: { + '#withContentSecurityPolicy':: d.fn(help='"Set the Content-Security-Policy header to HTTP responses. Unset if blank."', args=[d.arg(name='contentSecurityPolicy', type=d.T.string)]), + withContentSecurityPolicy(contentSecurityPolicy): { spec+: { web+: { httpConfig+: { headers+: { contentSecurityPolicy: contentSecurityPolicy } } } } }, + '#withStrictTransportSecurity':: d.fn(help='"Set the Strict-Transport-Security header to HTTP responses. Unset if blank. Please make sure that you use this with care as this header might force browsers to load Prometheus and the other applications hosted on the same domain and subdomains over HTTPS. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Strict-Transport-Security"', args=[d.arg(name='strictTransportSecurity', type=d.T.string)]), + withStrictTransportSecurity(strictTransportSecurity): { spec+: { web+: { httpConfig+: { headers+: { strictTransportSecurity: strictTransportSecurity } } } } }, + '#withXContentTypeOptions':: d.fn(help='"Set the X-Content-Type-Options header to HTTP responses. Unset if blank. Accepted value is nosniff. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Content-Type-Options"', args=[d.arg(name='xContentTypeOptions', type=d.T.string)]), + withXContentTypeOptions(xContentTypeOptions): { spec+: { web+: { httpConfig+: { headers+: { xContentTypeOptions: xContentTypeOptions } } } } }, + '#withXFrameOptions':: d.fn(help='"Set the X-Frame-Options header to HTTP responses. Unset if blank. Accepted values are deny and sameorigin. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options"', args=[d.arg(name='xFrameOptions', type=d.T.string)]), + withXFrameOptions(xFrameOptions): { spec+: { web+: { httpConfig+: { headers+: { xFrameOptions: xFrameOptions } } } } }, + '#withXXSSProtection':: d.fn(help='"Set the X-XSS-Protection header to all responses. Unset if blank. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-XSS-Protection"', args=[d.arg(name='xXSSProtection', type=d.T.string)]), + withXXSSProtection(xXSSProtection): { spec+: { web+: { httpConfig+: { headers+: { xXSSProtection: xXSSProtection } } } } }, + }, + '#withHttp2':: d.fn(help='"Enable HTTP/2 support. Note that HTTP/2 is only supported with TLS. When TLSConfig is not configured, HTTP/2 will be disabled. Whenever the value of the field changes, a rolling update will be triggered."', args=[d.arg(name='http2', type=d.T.boolean)]), + withHttp2(http2): { spec+: { web+: { httpConfig+: { http2: http2 } } } }, + }, + '#tlsConfig':: d.obj(help='"Defines the TLS parameters for HTTPS."'), + tlsConfig: { + '#cert':: d.obj(help='"Contains the TLS certificate for the server."'), + cert: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { cert+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { cert+: { secret+: { optional: optional } } } } } }, + }, + }, + '#client_ca':: d.obj(help='"Contains the CA certificate for client certificate authentication to the server."'), + client_ca: { + '#configMap':: d.obj(help='"ConfigMap containing data to use for the targets."'), + configMap: { + '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { client_ca+: { configMap+: { optional: optional } } } } } }, + }, + '#secret':: d.obj(help='"Secret containing data to use for the targets."'), + secret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { key: key } } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { name: name } } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { client_ca+: { secret+: { optional: optional } } } } } }, + }, + }, + '#keySecret':: d.obj(help='"Secret containing the TLS key for the server."'), + keySecret: { + '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]), + withKey(key): { spec+: { web+: { tlsConfig+: { keySecret+: { key: key } } } } }, + '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]), + withName(name): { spec+: { web+: { tlsConfig+: { keySecret+: { name: name } } } } }, + '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]), + withOptional(optional): { spec+: { web+: { tlsConfig+: { keySecret+: { optional: optional } } } } }, + }, + '#withCipherSuites':: d.fn(help='"List of supported cipher suites for TLS versions up to TLS 1.2. If empty, Go default cipher suites are used. Available cipher suites are documented in the go documentation: https://golang.org/pkg/crypto/tls/#pkg-constants"', args=[d.arg(name='cipherSuites', type=d.T.array)]), + withCipherSuites(cipherSuites): { spec+: { web+: { tlsConfig+: { cipherSuites: if std.isArray(v=cipherSuites) then cipherSuites else [cipherSuites] } } } }, + '#withCipherSuitesMixin':: d.fn(help='"List of supported cipher suites for TLS versions up to TLS 1.2. If empty, Go default cipher suites are used. Available cipher suites are documented in the go documentation: https://golang.org/pkg/crypto/tls/#pkg-constants"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='cipherSuites', type=d.T.array)]), + withCipherSuitesMixin(cipherSuites): { spec+: { web+: { tlsConfig+: { cipherSuites+: if std.isArray(v=cipherSuites) then cipherSuites else [cipherSuites] } } } }, + '#withClientAuthType':: d.fn(help='"Server policy for client authentication. Maps to ClientAuth Policies. For more detail on clientAuth options: https://golang.org/pkg/crypto/tls/#ClientAuthType"', args=[d.arg(name='clientAuthType', type=d.T.string)]), + withClientAuthType(clientAuthType): { spec+: { web+: { tlsConfig+: { clientAuthType: clientAuthType } } } }, + '#withCurvePreferences':: d.fn(help='"Elliptic curves that will be used in an ECDHE handshake, in preference order. Available curves are documented in the go documentation: https://golang.org/pkg/crypto/tls/#CurveID"', args=[d.arg(name='curvePreferences', type=d.T.array)]), + withCurvePreferences(curvePreferences): { spec+: { web+: { tlsConfig+: { curvePreferences: if std.isArray(v=curvePreferences) then curvePreferences else [curvePreferences] } } } }, + '#withCurvePreferencesMixin':: d.fn(help='"Elliptic curves that will be used in an ECDHE handshake, in preference order. Available curves are documented in the go documentation: https://golang.org/pkg/crypto/tls/#CurveID"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='curvePreferences', type=d.T.array)]), + withCurvePreferencesMixin(curvePreferences): { spec+: { web+: { tlsConfig+: { curvePreferences+: if std.isArray(v=curvePreferences) then curvePreferences else [curvePreferences] } } } }, + '#withMaxVersion':: d.fn(help='"Maximum TLS version that is acceptable. Defaults to TLS13."', args=[d.arg(name='maxVersion', type=d.T.string)]), + withMaxVersion(maxVersion): { spec+: { web+: { tlsConfig+: { maxVersion: maxVersion } } } }, + '#withMinVersion':: d.fn(help='"Minimum TLS version that is acceptable. Defaults to TLS12."', args=[d.arg(name='minVersion', type=d.T.string)]), + withMinVersion(minVersion): { spec+: { web+: { tlsConfig+: { minVersion: minVersion } } } }, + '#withPreferServerCipherSuites':: d.fn(help="\"Controls whether the server selects the client's most preferred cipher suite, or the server's most preferred cipher suite. If true then the server's preference, as expressed in the order of elements in cipherSuites, is used.\"", args=[d.arg(name='preferServerCipherSuites', type=d.T.boolean)]), + withPreferServerCipherSuites(preferServerCipherSuites): { spec+: { web+: { tlsConfig+: { preferServerCipherSuites: preferServerCipherSuites } } } }, + }, + '#withMaxConnections':: d.fn(help="\"Defines the maximum number of simultaneous connections A zero value means that Prometheus doesn't accept any incoming connection.\"", args=[d.arg(name='maxConnections', type=d.T.integer)]), + withMaxConnections(maxConnections): { spec+: { web+: { maxConnections: maxConnections } } }, + '#withPageTitle':: d.fn(help='"The prometheus web page title."', args=[d.arg(name='pageTitle', type=d.T.string)]), + withPageTitle(pageTitle): { spec+: { web+: { pageTitle: pageTitle } } }, + }, + '#withAdditionalArgs':: d.fn(help="\"AdditionalArgs allows setting additional arguments for the 'prometheus' container. \\n It is intended for e.g. activating hidden flags which are not supported by the dedicated configuration options yet. The arguments are passed as-is to the Prometheus container which may cause issues if they are invalid or not supported by the given Prometheus version. \\n In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged.\"", args=[d.arg(name='additionalArgs', type=d.T.array)]), + withAdditionalArgs(additionalArgs): { spec+: { additionalArgs: if std.isArray(v=additionalArgs) then additionalArgs else [additionalArgs] } }, + '#withAdditionalArgsMixin':: d.fn(help="\"AdditionalArgs allows setting additional arguments for the 'prometheus' container. \\n It is intended for e.g. activating hidden flags which are not supported by the dedicated configuration options yet. The arguments are passed as-is to the Prometheus container which may cause issues if they are invalid or not supported by the given Prometheus version. \\n In case of an argument conflict (e.g. an argument which is already set by the operator itself) or when providing an invalid argument, the reconciliation will fail and an error will be logged.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='additionalArgs', type=d.T.array)]), + withAdditionalArgsMixin(additionalArgs): { spec+: { additionalArgs+: if std.isArray(v=additionalArgs) then additionalArgs else [additionalArgs] } }, + '#withAllowOverlappingBlocks':: d.fn(help='"AllowOverlappingBlocks enables vertical compaction and vertical query merge in Prometheus. \\n *Deprecated: this flag has no effect for Prometheus >= 2.39.0 where overlapping blocks are enabled by default.*"', args=[d.arg(name='allowOverlappingBlocks', type=d.T.boolean)]), + withAllowOverlappingBlocks(allowOverlappingBlocks): { spec+: { allowOverlappingBlocks: allowOverlappingBlocks } }, + '#withBaseImage':: d.fn(help="\"*Deprecated: use 'spec.image' instead.*\"", args=[d.arg(name='baseImage', type=d.T.string)]), + withBaseImage(baseImage): { spec+: { baseImage: baseImage } }, + '#withBodySizeLimit':: d.fn(help='"BodySizeLimit defines per-scrape on response body size. Only valid in Prometheus versions 2.45.0 and newer."', args=[d.arg(name='bodySizeLimit', type=d.T.string)]), + withBodySizeLimit(bodySizeLimit): { spec+: { bodySizeLimit: bodySizeLimit } }, + '#withConfigMaps':: d.fn(help="\"ConfigMaps is a list of ConfigMaps in the same namespace as the Prometheus object, which shall be mounted into the Prometheus Pods. Each ConfigMap is added to the StatefulSet definition as a volume named `configmap-\u003cconfigmap-name\u003e`. The ConfigMaps are mounted into /etc/prometheus/configmaps/\u003cconfigmap-name\u003e in the 'prometheus' container.\"", args=[d.arg(name='configMaps', type=d.T.array)]), + withConfigMaps(configMaps): { spec+: { configMaps: if std.isArray(v=configMaps) then configMaps else [configMaps] } }, + '#withConfigMapsMixin':: d.fn(help="\"ConfigMaps is a list of ConfigMaps in the same namespace as the Prometheus object, which shall be mounted into the Prometheus Pods. Each ConfigMap is added to the StatefulSet definition as a volume named `configmap-\u003cconfigmap-name\u003e`. The ConfigMaps are mounted into /etc/prometheus/configmaps/\u003cconfigmap-name\u003e in the 'prometheus' container.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='configMaps', type=d.T.array)]), + withConfigMapsMixin(configMaps): { spec+: { configMaps+: if std.isArray(v=configMaps) then configMaps else [configMaps] } }, + '#withContainers':: d.fn(help='"Containers allows injecting additional containers or modifying operator generated containers. This can be used to allow adding an authentication proxy to the Pods or to change the behavior of an operator generated container. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. \\n The names of containers managed by the operator are: * `prometheus` * `config-reloader` * `thanos-sidecar` \\n Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."', args=[d.arg(name='containers', type=d.T.array)]), + withContainers(containers): { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } }, + '#withContainersMixin':: d.fn(help='"Containers allows injecting additional containers or modifying operator generated containers. This can be used to allow adding an authentication proxy to the Pods or to change the behavior of an operator generated container. Containers described here modify an operator generated container if they share the same name and modifications are done via a strategic merge patch. \\n The names of containers managed by the operator are: * `prometheus` * `config-reloader` * `thanos-sidecar` \\n Overriding containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]), + withContainersMixin(containers): { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } }, + '#withDisableCompaction':: d.fn(help='"When true, the Prometheus compaction is disabled."', args=[d.arg(name='disableCompaction', type=d.T.boolean)]), + withDisableCompaction(disableCompaction): { spec+: { disableCompaction: disableCompaction } }, + '#withEnableAdminAPI':: d.fn(help='"Enables access to the Prometheus web admin API. \\n WARNING: Enabling the admin APIs enables mutating endpoints, to delete data, shutdown Prometheus, and more. Enabling this should be done with care and the user is advised to add additional authentication authorization via a proxy to ensure only clients authorized to perform these actions can do so. \\n For more information: https://prometheus.io/docs/prometheus/latest/querying/api/#tsdb-admin-apis"', args=[d.arg(name='enableAdminAPI', type=d.T.boolean)]), + withEnableAdminAPI(enableAdminAPI): { spec+: { enableAdminAPI: enableAdminAPI } }, + '#withEnableFeatures':: d.fn(help='"Enable access to Prometheus feature flags. By default, no features are enabled. \\n Enabling features which are disabled by default is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice. \\n For more information see https://prometheus.io/docs/prometheus/latest/feature_flags/"', args=[d.arg(name='enableFeatures', type=d.T.array)]), + withEnableFeatures(enableFeatures): { spec+: { enableFeatures: if std.isArray(v=enableFeatures) then enableFeatures else [enableFeatures] } }, + '#withEnableFeaturesMixin':: d.fn(help='"Enable access to Prometheus feature flags. By default, no features are enabled. \\n Enabling features which are disabled by default is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice. \\n For more information see https://prometheus.io/docs/prometheus/latest/feature_flags/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='enableFeatures', type=d.T.array)]), + withEnableFeaturesMixin(enableFeatures): { spec+: { enableFeatures+: if std.isArray(v=enableFeatures) then enableFeatures else [enableFeatures] } }, + '#withEnableRemoteWriteReceiver':: d.fn(help='"Enable Prometheus to be used as a receiver for the Prometheus remote write protocol. \\n WARNING: This is not considered an efficient way of ingesting samples. Use it with caution for specific low-volume use cases. It is not suitable for replacing the ingestion via scraping and turning Prometheus into a push-based metrics collection system. For more information see https://prometheus.io/docs/prometheus/latest/querying/api/#remote-write-receiver \\n It requires Prometheus >= v2.33.0."', args=[d.arg(name='enableRemoteWriteReceiver', type=d.T.boolean)]), + withEnableRemoteWriteReceiver(enableRemoteWriteReceiver): { spec+: { enableRemoteWriteReceiver: enableRemoteWriteReceiver } }, + '#withEnforcedBodySizeLimit':: d.fn(help='"When defined, enforcedBodySizeLimit specifies a global limit on the size of uncompressed response body that will be accepted by Prometheus. Targets responding with a body larger than this many bytes will cause the scrape to fail. \\n It requires Prometheus >= v2.28.0."', args=[d.arg(name='enforcedBodySizeLimit', type=d.T.string)]), + withEnforcedBodySizeLimit(enforcedBodySizeLimit): { spec+: { enforcedBodySizeLimit: enforcedBodySizeLimit } }, + '#withEnforcedLabelLimit':: d.fn(help='"When defined, enforcedLabelLimit specifies a global limit on the number of labels per sample. The value overrides any `spec.labelLimit` set by ServiceMonitor, PodMonitor, Probe objects unless `spec.labelLimit` is greater than zero and less than `spec.enforcedLabelLimit`. \\n It requires Prometheus >= v2.27.0."', args=[d.arg(name='enforcedLabelLimit', type=d.T.integer)]), + withEnforcedLabelLimit(enforcedLabelLimit): { spec+: { enforcedLabelLimit: enforcedLabelLimit } }, + '#withEnforcedLabelNameLengthLimit':: d.fn(help='"When defined, enforcedLabelNameLengthLimit specifies a global limit on the length of labels name per sample. The value overrides any `spec.labelNameLengthLimit` set by ServiceMonitor, PodMonitor, Probe objects unless `spec.labelNameLengthLimit` is greater than zero and less than `spec.enforcedLabelNameLengthLimit`. \\n It requires Prometheus >= v2.27.0."', args=[d.arg(name='enforcedLabelNameLengthLimit', type=d.T.integer)]), + withEnforcedLabelNameLengthLimit(enforcedLabelNameLengthLimit): { spec+: { enforcedLabelNameLengthLimit: enforcedLabelNameLengthLimit } }, + '#withEnforcedLabelValueLengthLimit':: d.fn(help='"When not null, enforcedLabelValueLengthLimit defines a global limit on the length of labels value per sample. The value overrides any `spec.labelValueLengthLimit` set by ServiceMonitor, PodMonitor, Probe objects unless `spec.labelValueLengthLimit` is greater than zero and less than `spec.enforcedLabelValueLengthLimit`. \\n It requires Prometheus >= v2.27.0."', args=[d.arg(name='enforcedLabelValueLengthLimit', type=d.T.integer)]), + withEnforcedLabelValueLengthLimit(enforcedLabelValueLengthLimit): { spec+: { enforcedLabelValueLengthLimit: enforcedLabelValueLengthLimit } }, + '#withEnforcedNamespaceLabel':: d.fn(help="\"When not empty, a label will be added to \\n 1. All metrics scraped from `ServiceMonitor`, `PodMonitor`, `Probe` and `ScrapeConfig` objects. 2. All metrics generated from recording rules defined in `PrometheusRule` objects. 3. All alerts generated from alerting rules defined in `PrometheusRule` objects. 4. All vector selectors of PromQL expressions defined in `PrometheusRule` objects. \\n The label will not added for objects referenced in `spec.excludedFromEnforcement`. \\n The label's name is this field's value. The label's value is the namespace of the `ServiceMonitor`, `PodMonitor`, `Probe` or `PrometheusRule` object.\"", args=[d.arg(name='enforcedNamespaceLabel', type=d.T.string)]), + withEnforcedNamespaceLabel(enforcedNamespaceLabel): { spec+: { enforcedNamespaceLabel: enforcedNamespaceLabel } }, + '#withEnforcedSampleLimit':: d.fn(help='"When defined, enforcedSampleLimit specifies a global limit on the number of scraped samples that will be accepted. This overrides any `spec.sampleLimit` set by ServiceMonitor, PodMonitor, Probe objects unless `spec.sampleLimit` is greater than zero and less than than `spec.enforcedSampleLimit`. \\n It is meant to be used by admins to keep the overall number of samples/series under a desired limit."', args=[d.arg(name='enforcedSampleLimit', type=d.T.integer)]), + withEnforcedSampleLimit(enforcedSampleLimit): { spec+: { enforcedSampleLimit: enforcedSampleLimit } }, + '#withEnforcedTargetLimit':: d.fn(help='"When defined, enforcedTargetLimit specifies a global limit on the number of scraped targets. The value overrides any `spec.targetLimit` set by ServiceMonitor, PodMonitor, Probe objects unless `spec.targetLimit` is greater than zero and less than `spec.enforcedTargetLimit`. \\n It is meant to be used by admins to to keep the overall number of targets under a desired limit."', args=[d.arg(name='enforcedTargetLimit', type=d.T.integer)]), + withEnforcedTargetLimit(enforcedTargetLimit): { spec+: { enforcedTargetLimit: enforcedTargetLimit } }, + '#withEvaluationInterval':: d.fn(help='"Interval between rule evaluations. Default: \\"30s\\', args=[d.arg(name='evaluationInterval', type=d.T.string)]), + withEvaluationInterval(evaluationInterval): { spec+: { evaluationInterval: evaluationInterval } }, + '#withExcludedFromEnforcement':: d.fn(help='"List of references to PodMonitor, ServiceMonitor, Probe and PrometheusRule objects to be excluded from enforcing a namespace label of origin. \\n It is only applicable if `spec.enforcedNamespaceLabel` set to true."', args=[d.arg(name='excludedFromEnforcement', type=d.T.array)]), + withExcludedFromEnforcement(excludedFromEnforcement): { spec+: { excludedFromEnforcement: if std.isArray(v=excludedFromEnforcement) then excludedFromEnforcement else [excludedFromEnforcement] } }, + '#withExcludedFromEnforcementMixin':: d.fn(help='"List of references to PodMonitor, ServiceMonitor, Probe and PrometheusRule objects to be excluded from enforcing a namespace label of origin. \\n It is only applicable if `spec.enforcedNamespaceLabel` set to true."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='excludedFromEnforcement', type=d.T.array)]), + withExcludedFromEnforcementMixin(excludedFromEnforcement): { spec+: { excludedFromEnforcement+: if std.isArray(v=excludedFromEnforcement) then excludedFromEnforcement else [excludedFromEnforcement] } }, + '#withExternalLabels':: d.fn(help='"The labels to add to any time series or alerts when communicating with external systems (federation, remote storage, Alertmanager). Labels defined by `spec.replicaExternalLabelName` and `spec.prometheusExternalLabelName` take precedence over this list."', args=[d.arg(name='externalLabels', type=d.T.object)]), + withExternalLabels(externalLabels): { spec+: { externalLabels: externalLabels } }, + '#withExternalLabelsMixin':: d.fn(help='"The labels to add to any time series or alerts when communicating with external systems (federation, remote storage, Alertmanager). Labels defined by `spec.replicaExternalLabelName` and `spec.prometheusExternalLabelName` take precedence over this list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='externalLabels', type=d.T.object)]), + withExternalLabelsMixin(externalLabels): { spec+: { externalLabels+: externalLabels } }, + '#withExternalUrl':: d.fn(help='"The external URL under which the Prometheus service is externally available. This is necessary to generate correct URLs (for instance if Prometheus is accessible behind an Ingress resource)."', args=[d.arg(name='externalUrl', type=d.T.string)]), + withExternalUrl(externalUrl): { spec+: { externalUrl: externalUrl } }, + '#withHostAliases':: d.fn(help="\"Optional list of hosts and IPs that will be injected into the Pod's hosts file if specified.\"", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliases(hostAliases): { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostAliasesMixin':: d.fn(help="\"Optional list of hosts and IPs that will be injected into the Pod's hosts file if specified.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]), + withHostAliasesMixin(hostAliases): { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } }, + '#withHostNetwork':: d.fn(help="\"Use the host's network namespace if true. \\n Make sure to understand the security implications if you want to enable it (https://kubernetes.io/docs/concepts/configuration/overview/). \\n When hostNetwork is enabled, this will set the DNS policy to `ClusterFirstWithHostNet` automatically.\"", args=[d.arg(name='hostNetwork', type=d.T.boolean)]), + withHostNetwork(hostNetwork): { spec+: { hostNetwork: hostNetwork } }, + '#withIgnoreNamespaceSelectors':: d.fn(help='"When true, `spec.namespaceSelector` from all PodMonitor, ServiceMonitor and Probe objects will be ignored. They will only discover targets within the namespace of the PodMonitor, ServiceMonitor and Probe objec."', args=[d.arg(name='ignoreNamespaceSelectors', type=d.T.boolean)]), + withIgnoreNamespaceSelectors(ignoreNamespaceSelectors): { spec+: { ignoreNamespaceSelectors: ignoreNamespaceSelectors } }, + '#withImage':: d.fn(help='"Container image name for Prometheus. If specified, it takes precedence over the `spec.baseImage`, `spec.tag` and `spec.sha` fields. \\n Specifying `spec.version` is still necessary to ensure the Prometheus Operator knows which version of Prometheus is being configured. \\n If neither `spec.image` nor `spec.baseImage` are defined, the operator will use the latest upstream version of Prometheus available at the time when the operator was released."', args=[d.arg(name='image', type=d.T.string)]), + withImage(image): { spec+: { image: image } }, + '#withImagePullPolicy':: d.fn(help="\"Image pull policy for the 'prometheus', 'init-config-reloader' and 'config-reloader' containers. See https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy for more details.\"", args=[d.arg(name='imagePullPolicy', type=d.T.string)]), + withImagePullPolicy(imagePullPolicy): { spec+: { imagePullPolicy: imagePullPolicy } }, + '#withImagePullSecrets':: d.fn(help='"An optional list of references to Secrets in the same namespace to use for pulling images from registries. See http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecrets(imagePullSecrets): { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withImagePullSecretsMixin':: d.fn(help='"An optional list of references to Secrets in the same namespace to use for pulling images from registries. See http://kubernetes.io/docs/user-guide/images#specifying-imagepullsecrets-on-a-pod"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]), + withImagePullSecretsMixin(imagePullSecrets): { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } }, + '#withInitContainers':: d.fn(help='"InitContainers allows injecting initContainers to the Pod definition. Those can be used to e.g. fetch secrets for injection into the Prometheus configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. \\n The names of init container name managed by the operator are: * `init-config-reloader`. \\n Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainers(initContainers): { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withInitContainersMixin':: d.fn(help='"InitContainers allows injecting initContainers to the Pod definition. Those can be used to e.g. fetch secrets for injection into the Prometheus configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ InitContainers described here modify an operator generated init containers if they share the same name and modifications are done via a strategic merge patch. \\n The names of init container name managed by the operator are: * `init-config-reloader`. \\n Overriding init containers is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]), + withInitContainersMixin(initContainers): { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } }, + '#withLabelLimit':: d.fn(help='"Per-scrape limit on number of labels that will be accepted for a sample. Only valid in Prometheus versions 2.45.0 and newer."', args=[d.arg(name='labelLimit', type=d.T.integer)]), + withLabelLimit(labelLimit): { spec+: { labelLimit: labelLimit } }, + '#withLabelNameLengthLimit':: d.fn(help='"Per-scrape limit on length of labels name that will be accepted for a sample. Only valid in Prometheus versions 2.45.0 and newer."', args=[d.arg(name='labelNameLengthLimit', type=d.T.integer)]), + withLabelNameLengthLimit(labelNameLengthLimit): { spec+: { labelNameLengthLimit: labelNameLengthLimit } }, + '#withLabelValueLengthLimit':: d.fn(help='"Per-scrape limit on length of labels value that will be accepted for a sample. Only valid in Prometheus versions 2.45.0 and newer."', args=[d.arg(name='labelValueLengthLimit', type=d.T.integer)]), + withLabelValueLengthLimit(labelValueLengthLimit): { spec+: { labelValueLengthLimit: labelValueLengthLimit } }, + '#withListenLocal':: d.fn(help="\"When true, the Prometheus server listens on the loopback address instead of the Pod IP's address.\"", args=[d.arg(name='listenLocal', type=d.T.boolean)]), + withListenLocal(listenLocal): { spec+: { listenLocal: listenLocal } }, + '#withLogFormat':: d.fn(help='"Log format for Log level for Prometheus and the config-reloader sidecar."', args=[d.arg(name='logFormat', type=d.T.string)]), + withLogFormat(logFormat): { spec+: { logFormat: logFormat } }, + '#withLogLevel':: d.fn(help='"Log level for Prometheus and the config-reloader sidecar."', args=[d.arg(name='logLevel', type=d.T.string)]), + withLogLevel(logLevel): { spec+: { logLevel: logLevel } }, + '#withMinReadySeconds':: d.fn(help='"Minimum number of seconds for which a newly created Pod should be ready without any of its container crashing for it to be considered available. Defaults to 0 (pod will be considered available as soon as it is ready) \\n This is an alpha field from kubernetes 1.22 until 1.24 which requires enabling the StatefulSetMinReadySeconds feature gate."', args=[d.arg(name='minReadySeconds', type=d.T.integer)]), + withMinReadySeconds(minReadySeconds): { spec+: { minReadySeconds: minReadySeconds } }, + '#withNodeSelector':: d.fn(help='"Defines on which Nodes the Pods are scheduled."', args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelector(nodeSelector): { spec+: { nodeSelector: nodeSelector } }, + '#withNodeSelectorMixin':: d.fn(help='"Defines on which Nodes the Pods are scheduled."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelector', type=d.T.object)]), + withNodeSelectorMixin(nodeSelector): { spec+: { nodeSelector+: nodeSelector } }, + '#withOverrideHonorLabels':: d.fn(help='"When true, Prometheus resolves label conflicts by renaming the labels in the scraped data to \\"exported_