content large_stringlengths 3 20.5k | url large_stringlengths 54 193 | branch large_stringclasses 4
values | source large_stringclasses 42
values | embeddings listlengths 384 384 | score float64 -0.21 0.65 |
|---|---|---|---|---|---|
be used as follows:: In [1]: import numpy as np In [2]: %memit np.zeros(1e7) maximum of 3: 76.402344 MB per loop For more details, see the docstrings of the magics, using ``%memit?`` and ``%mprun?``. Using Cython ============ If profiling of the Python code reveals that the Python interpreter overhead is larger by one order of magnitude or more than the cost of the actual numerical computation (e.g. ``for`` loops over vector components, nested evaluation of conditional expression, scalar arithmetic...), it is probably adequate to extract the hotspot portion of the code as a standalone function in a ``.pyx`` file, add static type declarations and then use Cython to generate a C program suitable to be compiled as a Python extension module. The `Cython's documentation `\_ contains a tutorial and reference guide for developing such a module. For more information about developing in Cython for scikit-learn, see :ref:`cython`. .. \_profiling-compiled-extension: Profiling compiled extensions ============================= When working with compiled extensions (written in C/C++ with a wrapper or directly as Cython extension), the default Python profiler is useless: we need a dedicated tool to introspect what's happening inside the compiled extension itself. Using yep and gperftools ------------------------ Easy profiling without special compilation options use yep: - https://pypi.org/project/yep/ - https://fa.bianp.net/blog/2011/a-profiler-for-python-extensions Using a debugger, gdb --------------------- \* It is helpful to use ``gdb`` to debug. In order to do so, one must use a Python interpreter built with debug support (debug symbols and proper optimization). To create a new conda environment (which you might need to deactivate and reactivate after building/installing) with a source-built CPython interpreter: .. code-block:: bash git clone https://github.com/python/cpython.git conda create -n debug-scikit-dev conda activate debug-scikit-dev cd cpython mkdir debug cd debug ../configure --prefix=$CONDA\_PREFIX --with-pydebug make EXTRA\_CFLAGS='-DPy\_DEBUG' -j make install Using gprof ----------- In order to profile compiled Python extensions one could use ``gprof`` after having recompiled the project with ``gcc -pg`` and using the ``python-dbg`` variant of the interpreter on debian / ubuntu: however this approach requires to also have ``numpy`` and ``scipy`` recompiled with ``-pg`` which is rather complicated to get working. Fortunately there exist two alternative profilers that don't require you to recompile everything. Using valgrind / callgrind / kcachegrind ---------------------------------------- kcachegrind ~~~~~~~~~~~ ``yep`` can be used to create a profiling report. ``kcachegrind`` provides a graphical environment to visualize this report: .. prompt:: bash $ # Run yep to profile some python script python -m yep -c my\_file.py .. prompt:: bash $ # open my\_file.py.callgrin with kcachegrind kcachegrind my\_file.py.prof .. note:: ``yep`` can be executed with the argument ``--lines`` or ``-l`` to compile a profiling report 'line by line'. Multi-core parallelism using ``joblib.Parallel`` ================================================ See `joblib documentation `\_ .. \_warm-restarts: A simple algorithmic trick: warm restarts ========================================= See the glossary entry for :term:`warm\_start` | https://github.com/scikit-learn/scikit-learn/blob/main//doc/developers/performance.rst | main | scikit-learn | [
-0.022180281579494476,
0.01745724119246006,
-0.08538910746574402,
0.0026375306770205498,
0.040862154215574265,
-0.16128666698932648,
0.014445065520703793,
0.050177887082099915,
-0.07238749414682388,
0.01686561107635498,
-0.02919928915798664,
-0.020290805026888847,
0.008398523554205894,
-0.... | 0.074608 |
.. AUTO-GENERATED. Please DO NOT edit manually. .. role:: raw-html-m2r(raw) :format: html .. list-table:: :header-rows: 1 \* - :spelling:ignore:`Key` - Description - Type - Default \* - :spelling:ignore:`MTU` - Configure the underlying network MTU to overwrite auto-detected MTU. This value doesn't change the host network interface MTU i.e. eth0 or ens0. It changes the MTU for cilium\_net@cilium\_host, cilium\_host@cilium\_net, cilium\_vxlan and lxc\_health interfaces. - int - ``0`` \* - :spelling:ignore:`affinity` - Affinity for cilium-agent. - object - ``{"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`agent` - Install the cilium agent resources. - bool - ``true`` \* - :spelling:ignore:`agentNotReadyTaintKey` - Configure the key of the taint indicating that Cilium is not ready on the node. When set to a value starting with ``ignore-taint.cluster-autoscaler.kubernetes.io/``\ , the Cluster Autoscaler will ignore the taint on its decisions, allowing the cluster to scale up. - string - ``"node.cilium.io/agent-not-ready"`` \* - :spelling:ignore:`aksbyocni.enabled` - Enable AKS BYOCNI integration. Note that this is incompatible with AKS clusters not created in BYOCNI mode: use Azure integration (\ ``azure.enabled``\ ) instead. - bool - ``false`` \* - :spelling:ignore:`alibabacloud.enabled` - Enable AlibabaCloud ENI integration - bool - ``false`` \* - :spelling:ignore:`alibabacloud.nodeSpec.securityGroupTags` - - list - ``[]`` \* - :spelling:ignore:`alibabacloud.nodeSpec.securityGroups` - - list - ``[]`` \* - :spelling:ignore:`alibabacloud.nodeSpec.vSwitchTags` - - list - ``[]`` \* - :spelling:ignore:`alibabacloud.nodeSpec.vSwitches` - - list - ``[]`` \* - :spelling:ignore:`annotateK8sNode` - Annotate k8s node upon initialization with Cilium's metadata. - bool - ``false`` \* - :spelling:ignore:`annotations` - Annotations to be added to all top-level cilium-agent objects (resources under templates/cilium-agent) - object - ``{}`` \* - :spelling:ignore:`apiRateLimit` - The api-rate-limit option can be used to overwrite individual settings of the default configuration for rate limiting calls to the Cilium Agent API - string - ``nil`` \* - :spelling:ignore:`authentication.enabled` - Enable authentication processing and garbage collection. Note that if disabled, policy enforcement will still block requests that require authentication. But the resulting authentication requests for these requests will not be processed, therefore the requests not be allowed. - bool - ``true`` \* - :spelling:ignore:`authentication.gcInterval` - Interval for garbage collection of auth map entries. - string - ``"5m0s"`` \* - :spelling:ignore:`authentication.mutual.connectTimeout` - Timeout for connecting to the remote node TCP socket - string - ``"5s"`` \* - :spelling:ignore:`authentication.mutual.port` - Port on the agent where mutual authentication handshakes between agents will be performed - int - ``4250`` \* - :spelling:ignore:`authentication.mutual.spire.adminSocketPath` - SPIRE socket path where the SPIRE delegated api agent is listening - string - ``"/run/spire/sockets/admin.sock"`` \* - :spelling:ignore:`authentication.mutual.spire.agentSocketPath` - SPIRE socket path where the SPIRE workload agent is listening. Applies to both the Cilium Agent and Operator - string - ``"/run/spire/sockets/agent/agent.sock"`` \* - :spelling:ignore:`authentication.mutual.spire.annotations` - Annotations to be added to all top-level spire objects (resources under templates/spire) - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.connectionTimeout` - SPIRE connection timeout - string - ``"30s"`` \* - :spelling:ignore:`authentication.mutual.spire.enabled` - Enable SPIRE integration (beta) - bool - ``false`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.affinity` - SPIRE agent affinity configuration - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.annotations` - SPIRE agent annotations - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.image` - SPIRE agent image - object - ``{"digest":"sha256:7bb8469af28f00908a96a8a51fb36cbbcdfb85efe78e94b492f949f33284bf4d","override":null,"pullPolicy":"Always","repository":"ghcr.io/spiffe/spire-agent","tag":"1.14.1","useDigest":true}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.labels` - SPIRE agent labels - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.nodeSelector` - SPIRE agent nodeSelector configuration ref: ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.podSecurityContext` - Security context to be added to spire agent pods. SecurityContext holds pod-level security attributes and common container settings. ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/#set-the-security-context-for-a-pod - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.priorityClassName` - The priority class to use for the spire agent - string - ``""`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.resources` - container resource limits & requests - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.securityContext` - Security context to be added to spire agent containers. SecurityContext | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.03811325132846832,
0.019189581274986267,
0.00567285530269146,
0.00796604249626398,
-0.016547435894608498,
-0.05699606612324715,
-0.08139622956514359,
-0.003094759304076433,
-0.050929293036460876,
-0.055424511432647705,
0.023138314485549927,
-0.10842844843864441,
0.06734510511159897,
-0.... | 0.103989 |
ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/#set-the-security-context-for-a-pod - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.priorityClassName` - The priority class to use for the spire agent - string - ``""`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.resources` - container resource limits & requests - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.securityContext` - Security context to be added to spire agent containers. SecurityContext holds pod-level security attributes and common container settings. ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/#set-the-security-context-for-a-container - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.serviceAccount` - SPIRE agent service account - object - ``{"create":true,"name":"spire-agent"}`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.skipKubeletVerification` - SPIRE Workload Attestor kubelet verification. - bool - ``true`` \* - :spelling:ignore:`authentication.mutual.spire.install.agent.tolerations` - SPIRE agent tolerations configuration By default it follows the same tolerations as the agent itself to allow the Cilium agent on this node to connect to SPIRE. ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[{"effect":"NoSchedule","key":"node.kubernetes.io/not-ready"},{"effect":"NoSchedule","key":"node-role.kubernetes.io/master"},{"effect":"NoSchedule","key":"node-role.kubernetes.io/control-plane"},{"effect":"NoSchedule","key":"node.cloudprovider.kubernetes.io/uninitialized","value":"true"},{"key":"CriticalAddonsOnly","operator":"Exists"}]`` \* - :spelling:ignore:`authentication.mutual.spire.install.enabled` - Enable SPIRE installation. This will only take effect only if authentication.mutual.spire.enabled is true - bool - ``true`` \* - :spelling:ignore:`authentication.mutual.spire.install.existingNamespace` - SPIRE namespace already exists. Set to true if Helm should not create, manage, and import the SPIRE namespace. - bool - ``false`` \* - :spelling:ignore:`authentication.mutual.spire.install.initImage` - init container image of SPIRE agent and server - object - ``{"digest":"sha256:e226d6308690dbe282443c8c7e57365c96b5228f0fe7f40731b5d84d37a06839","override":null,"pullPolicy":"Always","repository":"docker.io/library/busybox","tag":"1.37.0","useDigest":true}`` \* - :spelling:ignore:`authentication.mutual.spire.install.namespace` - SPIRE namespace to install into - string - ``"cilium-spire"`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.affinity` - SPIRE server affinity configuration - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.annotations` - SPIRE server annotations - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.ca.keyType` - SPIRE CA key type AWS requires the use of RSA. EC cryptography is not supported - string - ``"rsa-4096"`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.ca.subject` - SPIRE CA Subject - object - ``{"commonName":"Cilium SPIRE CA","country":"US","organization":"SPIRE"}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.dataStorage.accessMode` - Access mode of the SPIRE server data storage - string - ``"ReadWriteOnce"`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.dataStorage.enabled` - Enable SPIRE server data storage - bool - ``true`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.dataStorage.size` - Size of the SPIRE server data storage - string - ``"1Gi"`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.dataStorage.storageClass` - StorageClass of the SPIRE server data storage - string - ``nil`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.image` - SPIRE server image - object - ``{"digest":"sha256:46a740705d5e15839552b1307aff44ef5ac42d9b444d073b4ccefd87c5269283","override":null,"pullPolicy":"Always","repository":"ghcr.io/spiffe/spire-server","tag":"1.14.1","useDigest":true}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.initContainers` - SPIRE server init containers - list - ``[]`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.labels` - SPIRE server labels - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.nodeSelector` - SPIRE server nodeSelector configuration ref: ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.podSecurityContext` - Security context to be added to spire server pods. SecurityContext holds pod-level security attributes and common container settings. ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/#set-the-security-context-for-a-pod - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.priorityClassName` - The priority class to use for the spire server - string - ``""`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.resources` - container resource limits & requests - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.securityContext` - Security context to be added to spire server containers. SecurityContext holds pod-level security attributes and common container settings. ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/#set-the-security-context-for-a-container - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.service.annotations` - Annotations to be added to the SPIRE server service - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.service.labels` - Labels to be added to the SPIRE server service - object - ``{}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.service.type` - Service type for the SPIRE server service - string - ``"ClusterIP"`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.serviceAccount` - SPIRE server service account - object - ``{"create":true,"name":"spire-server"}`` \* - :spelling:ignore:`authentication.mutual.spire.install.server.tolerations` - SPIRE server tolerations configuration ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`authentication.mutual.spire.serverAddress` - SPIRE server address used by Cilium Operator If k8s Service DNS along with port number is used (e.g. :raw-html-m2r:``.\ :raw-html-m2r:``.svc(.\*):\ :raw-html-m2r:`` format), Cilium Operator will resolve its address by looking up the clusterIP from Service resource. Example values: 10.0.0.1:8081, spire-server.cilium-spire.svc:8081 - string - ``nil`` \* - :spelling:ignore:`authentication.mutual.spire.trustDomain` - SPIFFE trust domain to use for fetching certificates - string - ``"spiffe.cilium"`` \* - :spelling:ignore:`authentication.queueSize` - Buffer size | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.017714975401759148,
0.07418932765722275,
-0.021162835881114006,
-0.002539504086598754,
0.014043361879885197,
0.017742693424224854,
0.0603921078145504,
-0.017666341736912727,
0.0038276943378150463,
0.06567282229661942,
-0.01743379607796669,
-0.04892820119857788,
0.013179179280996323,
0.05... | 0.086442 |
used (e.g. :raw-html-m2r:``.\ :raw-html-m2r:``.svc(.\*):\ :raw-html-m2r:`` format), Cilium Operator will resolve its address by looking up the clusterIP from Service resource. Example values: 10.0.0.1:8081, spire-server.cilium-spire.svc:8081 - string - ``nil`` \* - :spelling:ignore:`authentication.mutual.spire.trustDomain` - SPIFFE trust domain to use for fetching certificates - string - ``"spiffe.cilium"`` \* - :spelling:ignore:`authentication.queueSize` - Buffer size of the channel Cilium uses to receive authentication events from the signal map. - int - ``1024`` \* - :spelling:ignore:`authentication.rotatedIdentitiesQueueSize` - Buffer size of the channel Cilium uses to receive certificate expiration events from auth handlers. - int - ``1024`` \* - :spelling:ignore:`autoDirectNodeRoutes` - Enable installation of PodCIDR routes between worker nodes if worker nodes share a common L2 network segment. - bool - ``false`` \* - :spelling:ignore:`azure.enabled` - Enable Azure integration. Note that this is incompatible with AKS clusters created in BYOCNI mode: use AKS BYOCNI integration (\ ``aksbyocni.enabled``\ ) instead. - bool - ``false`` \* - :spelling:ignore:`azure.nodeSpec.azureInterfaceName` - - string - ``""`` \* - :spelling:ignore:`bandwidthManager` - Enable bandwidth manager to optimize TCP and UDP workloads and allow for rate-limiting traffic from individual Pods with EDT (Earliest Departure Time) through the "kubernetes.io/egress-bandwidth" Pod annotation. - object - ``{"bbr":false,"bbrHostNamespaceOnly":false,"enabled":false}`` \* - :spelling:ignore:`bandwidthManager.bbr` - Activate BBR TCP congestion control for Pods - bool - ``false`` \* - :spelling:ignore:`bandwidthManager.bbrHostNamespaceOnly` - Activate BBR TCP congestion control for Pods in the host namespace only. - bool - ``false`` \* - :spelling:ignore:`bandwidthManager.enabled` - Enable bandwidth manager infrastructure (also prerequirement for BBR) - bool - ``false`` \* - :spelling:ignore:`bgpControlPlane` - This feature set enables virtual BGP routers to be created via BGP CRDs. - object - ``{"enabled":false,"legacyOriginAttribute":{"enabled":false},"routerIDAllocation":{"ipPool":"","mode":"default"},"secretsNamespace":{"create":false,"name":"kube-system"},"statusReport":{"enabled":true}}`` \* - :spelling:ignore:`bgpControlPlane.enabled` - Enables the BGP control plane. - bool - ``false`` \* - :spelling:ignore:`bgpControlPlane.legacyOriginAttribute` - Legacy BGP ORIGIN attribute settings - object - ``{"enabled":false}`` \* - :spelling:ignore:`bgpControlPlane.legacyOriginAttribute.enabled` - Enable/Disable advertising LoadBalancerIP routes with the legacy BGP ORIGIN attribute value INCOMPLETE (2) instead of the default IGP (0). Enable for compatibility with the legacy behavior of MetalLB integration. - bool - ``false`` \* - :spelling:ignore:`bgpControlPlane.routerIDAllocation` - BGP router-id allocation mode - object - ``{"ipPool":"","mode":"default"}`` \* - :spelling:ignore:`bgpControlPlane.routerIDAllocation.ipPool` - IP pool to allocate the BGP router-id from when the mode is ip-pool. - string - ``""`` \* - :spelling:ignore:`bgpControlPlane.routerIDAllocation.mode` - BGP router-id allocation mode. In default mode, the router-id is derived from the IPv4 address if it is available, or else it is determined by the lower 32 bits of the MAC address. - string - ``"default"`` \* - :spelling:ignore:`bgpControlPlane.secretsNamespace` - SecretsNamespace is the namespace which BGP support will retrieve secrets from. - object - ``{"create":false,"name":"kube-system"}`` \* - :spelling:ignore:`bgpControlPlane.secretsNamespace.create` - Create secrets namespace for BGP secrets. - bool - ``false`` \* - :spelling:ignore:`bgpControlPlane.secretsNamespace.name` - The name of the secret namespace to which Cilium agents are given read access - string - ``"kube-system"`` \* - :spelling:ignore:`bgpControlPlane.statusReport` - Status reporting settings - object - ``{"enabled":true}`` \* - :spelling:ignore:`bgpControlPlane.statusReport.enabled` - Enable/Disable BGP status reporting It is recommended to enable status reporting in general, but if you have any issue such as high API server load, you can disable it by setting this to false. - bool - ``true`` \* - :spelling:ignore:`bpf.authMapMax` - Configure the maximum number of entries in auth map. - int - ``524288`` \* - :spelling:ignore:`bpf.autoMount.enabled` - Enable automatic mount of BPF filesystem When ``autoMount`` is enabled, the BPF filesystem is mounted at ``bpf.root`` path on the underlying host and inside the cilium agent pod. If users disable ``autoMount``\ , it's expected that users have mounted bpffs filesystem at the specified ``bpf.root`` volume, and then the volume will be mounted inside the cilium agent pod at the same path. - bool - ``true`` \* - :spelling:ignore:`bpf.ctAccounting` - Enable CT accounting | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.04875454679131508,
-0.020800625905394554,
-0.11579687893390656,
-0.01811503991484642,
-0.03127197176218033,
-0.09113752096891403,
-0.04152808338403702,
0.0009209308191202581,
0.025172794237732887,
-0.07170075923204422,
0.03192012757062912,
-0.07394677400588989,
0.012689833529293537,
-0.... | 0.125373 |
inside the cilium agent pod. If users disable ``autoMount``\ , it's expected that users have mounted bpffs filesystem at the specified ``bpf.root`` volume, and then the volume will be mounted inside the cilium agent pod at the same path. - bool - ``true`` \* - :spelling:ignore:`bpf.ctAccounting` - Enable CT accounting for packets and bytes - bool - ``false`` \* - :spelling:ignore:`bpf.ctAnyMax` - Configure the maximum number of entries for the non-TCP connection tracking table. - int - ``262144`` \* - :spelling:ignore:`bpf.ctTcpMax` - Configure the maximum number of entries in the TCP connection tracking table. - int - ``524288`` \* - :spelling:ignore:`bpf.datapathMode` - Mode for Pod devices for the core datapath (auto, veth, netkit, netkit-l2) - string - ``veth`` \* - :spelling:ignore:`bpf.disableExternalIPMitigation` - Disable ExternalIP mitigation (CVE-2020-8554) - bool - ``false`` \* - :spelling:ignore:`bpf.distributedLRU` - Control to use a distributed per-CPU backend memory for the core BPF LRU maps which Cilium uses. This improves performance significantly, but it is also recommended to increase BPF map sizing along with that. - object - ``{"enabled":false}`` \* - :spelling:ignore:`bpf.distributedLRU.enabled` - Enable distributed LRU backend memory. For compatibility with existing installations it is off by default. - bool - ``false`` \* - :spelling:ignore:`bpf.enableTCX` - Attach endpoint programs using tcx instead of legacy tc hooks on supported kernels. - bool - ``true`` \* - :spelling:ignore:`bpf.events` - Control events generated by the Cilium datapath exposed to Cilium monitor and Hubble. Helm configuration for BPF events map rate limiting is experimental and might change in upcoming releases. - object - ``{"default":{"burstLimit":null,"rateLimit":null},"drop":{"enabled":true},"policyVerdict":{"enabled":true},"trace":{"enabled":true}}`` \* - :spelling:ignore:`bpf.events.default` - Default settings for all types of events except dbg. - object - ``{"burstLimit":null,"rateLimit":null}`` \* - :spelling:ignore:`bpf.events.default.burstLimit` - Configure the maximum number of messages that can be written to BPF events map in 1 second. If burstLimit is greater than 0, non-zero value for rateLimit must also be provided lest the configuration is considered invalid. Setting both burstLimit and rateLimit to 0 disables BPF events rate limiting. - int - ``0`` \* - :spelling:ignore:`bpf.events.default.rateLimit` - Configure the limit of messages per second that can be written to BPF events map. The number of messages is averaged, meaning that if no messages were written to the map over 5 seconds, it's possible to write more events in the 6th second. If rateLimit is greater than 0, non-zero value for burstLimit must also be provided lest the configuration is considered invalid. Setting both burstLimit and rateLimit to 0 disables BPF events rate limiting. - int - ``0`` \* - :spelling:ignore:`bpf.events.drop.enabled` - Enable drop events. - bool - ``true`` \* - :spelling:ignore:`bpf.events.policyVerdict.enabled` - Enable policy verdict events. - bool - ``true`` \* - :spelling:ignore:`bpf.events.trace.enabled` - Enable trace events. - bool - ``true`` \* - :spelling:ignore:`bpf.hostLegacyRouting` - Configure whether direct routing mode should route traffic via host stack (true) or directly and more efficiently out of BPF (false) if the kernel supports it. The latter has the implication that it will also bypass netfilter in the host namespace. - bool - ``false`` \* - :spelling:ignore:`bpf.lbAlgorithmAnnotation` - Enable the option to define the load balancing algorithm on a per-service basis through service.cilium.io/lb-algorithm annotation. - bool - ``false`` \* - :spelling:ignore:`bpf.lbExternalClusterIP` - Allow cluster external access to ClusterIP services. - bool - ``false`` \* - :spelling:ignore:`bpf.lbMapMax` - Configure the maximum number of service entries in the load balancer maps. - int - ``65536`` \* - :spelling:ignore:`bpf.lbModeAnnotation` - Enable the option to define the load balancing mode (SNAT or DSR) on a per-service basis through service.cilium.io/forwarding-mode annotation. - bool - ``false`` \* - :spelling:ignore:`bpf.lbSourceRangeAllTypes` - Enable loadBalancerSourceRanges CIDR filtering for all service types, not just LoadBalancer services. The corresponding NodePort and | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.04333018139004707,
-0.03940238803625107,
-0.07924824953079224,
-0.07370056211948395,
-0.02578354999423027,
-0.06031276658177376,
0.023451006039977074,
0.05428481101989746,
0.029179319739341736,
-0.0006332012126222253,
0.05631432309746742,
-0.08592195063829422,
0.04158616065979004,
-0.010... | 0.168139 |
int - ``65536`` \* - :spelling:ignore:`bpf.lbModeAnnotation` - Enable the option to define the load balancing mode (SNAT or DSR) on a per-service basis through service.cilium.io/forwarding-mode annotation. - bool - ``false`` \* - :spelling:ignore:`bpf.lbSourceRangeAllTypes` - Enable loadBalancerSourceRanges CIDR filtering for all service types, not just LoadBalancer services. The corresponding NodePort and ClusterIP (if enabled for cluster-external traffic) will also apply the CIDR filter. - bool - ``false`` \* - :spelling:ignore:`bpf.mapDynamicSizeRatio` - Configure auto-sizing for all BPF maps based on available memory. ref: https://docs.cilium.io/en/stable/network/ebpf/maps/ - float64 - ``0.0025`` \* - :spelling:ignore:`bpf.masquerade` - Enable native IP masquerade support in eBPF - bool - ``false`` \* - :spelling:ignore:`bpf.monitorAggregation` - Configure the level of aggregation for monitor notifications. Valid options are none, low, medium, maximum. - string - ``"medium"`` \* - :spelling:ignore:`bpf.monitorFlags` - Configure which TCP flags trigger notifications when seen for the first time in a connection. - string - ``"all"`` \* - :spelling:ignore:`bpf.monitorInterval` - Configure the typical time between monitor notifications for active connections. - string - ``"5s"`` \* - :spelling:ignore:`bpf.monitorTraceIPOption` - Configure the IP tracing option type. This option is used to specify the IP option type to use for tracing. The value must be an integer between 0 and 255. @schema type: [null, integer] minimum: 0 maximum: 255 @schema - int - ``0`` \* - :spelling:ignore:`bpf.natMax` - Configure the maximum number of entries for the NAT table. - int - ``524288`` \* - :spelling:ignore:`bpf.neighMax` - Configure the maximum number of entries for the neighbor table. - int - ``524288`` \* - :spelling:ignore:`bpf.nodeMapMax` - Configures the maximum number of entries for the node table. - int - ``nil`` \* - :spelling:ignore:`bpf.policyMapMax` - Configure the maximum number of entries in endpoint policy map (per endpoint). @schema type: [null, integer] @schema - int - ``16384`` \* - :spelling:ignore:`bpf.policyMapPressureMetricsThreshold` - Configure threshold for emitting pressure metrics of policy maps. @schema type: [null, number] @schema - float64 - ``0.1`` \* - :spelling:ignore:`bpf.policyStatsMapMax` - Configure the maximum number of entries in global policy stats map. @schema type: [null, integer] @schema - int - ``65536`` \* - :spelling:ignore:`bpf.preallocateMaps` - Enables pre-allocation of eBPF map values. This increases memory usage but can reduce latency. - bool - ``false`` \* - :spelling:ignore:`bpf.root` - Configure the mount point for the BPF filesystem - string - ``"/sys/fs/bpf"`` \* - :spelling:ignore:`bpf.tproxy` - Configure the eBPF-based TPROXY (beta) to reduce reliance on iptables rules for implementing Layer 7 policy. - bool - ``false`` \* - :spelling:ignore:`bpf.vlanBypass` - Configure explicitly allowed VLAN id's for bpf logic bypass. [0] will allow all VLAN id's without any filtering. - list - ``[]`` \* - :spelling:ignore:`bpfClockProbe` - Enable BPF clock source probing for more efficient tick retrieval. - bool - ``false`` \* - :spelling:ignore:`certgen` - Configure certificate generation for Hubble integration. If hubble.tls.auto.method=cronJob, these values are used for the Kubernetes CronJob which will be scheduled regularly to (re)generate any certificates not provided manually. - object - ``{"affinity":{},"annotations":{"cronJob":{},"job":{}},"cronJob":{"failedJobsHistoryLimit":1,"successfulJobsHistoryLimit":3},"extraVolumeMounts":[],"extraVolumes":[],"generateCA":true,"image":{"digest":"sha256:19921f48ee7e2295ea4dca955878a6cd8d70e6d4219d08f688e866ece9d95d4d","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/certgen","tag":"v0.3.2","useDigest":true},"nodeSelector":{},"podLabels":{},"priorityClassName":"","resources":{},"tolerations":[],"ttlSecondsAfterFinished":null}`` \* - :spelling:ignore:`certgen.affinity` - Affinity for certgen - object - ``{}`` \* - :spelling:ignore:`certgen.annotations` - Annotations to be added to the hubble-certgen initial Job and CronJob - object - ``{"cronJob":{},"job":{}}`` \* - :spelling:ignore:`certgen.cronJob.failedJobsHistoryLimit` - The number of failed finished jobs to keep - int - ``1`` \* - :spelling:ignore:`certgen.cronJob.successfulJobsHistoryLimit` - The number of successful finished jobs to keep - int - ``3`` \* - :spelling:ignore:`certgen.extraVolumeMounts` - Additional certgen volumeMounts. - list - ``[]`` \* - :spelling:ignore:`certgen.extraVolumes` - Additional certgen volumes. - list - ``[]`` \* - :spelling:ignore:`certgen.generateCA` - When set to true the certificate authority secret is created. - bool - ``true`` \* - :spelling:ignore:`certgen.nodeSelector` - Node selector for certgen ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{}`` \* - :spelling:ignore:`certgen.podLabels` - Labels | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.02383640594780445,
0.014108838513493538,
-0.08891110122203827,
-0.031310271471738815,
-0.0028200424276292324,
0.0010377926519140601,
-0.026199964806437492,
0.011439451947808266,
-0.05967992544174194,
-0.03362512215971947,
0.011562828905880451,
-0.13147549331188202,
0.0009048490319401026,
... | 0.169762 |
- ``[]`` \* - :spelling:ignore:`certgen.extraVolumes` - Additional certgen volumes. - list - ``[]`` \* - :spelling:ignore:`certgen.generateCA` - When set to true the certificate authority secret is created. - bool - ``true`` \* - :spelling:ignore:`certgen.nodeSelector` - Node selector for certgen ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{}`` \* - :spelling:ignore:`certgen.podLabels` - Labels to be added to hubble-certgen pods - object - ``{}`` \* - :spelling:ignore:`certgen.priorityClassName` - Priority class for certgen ref: https://kubernetes.io/docs/concepts/scheduling-eviction/pod-priority-preemption/#priorityclass - string - ``""`` \* - :spelling:ignore:`certgen.resources` - Resource limits for certgen ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers - object - ``{}`` \* - :spelling:ignore:`certgen.tolerations` - Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`certgen.ttlSecondsAfterFinished` - Seconds after which the completed job pod will be deleted - string - ``nil`` \* - :spelling:ignore:`cgroup` - Configure cgroup related configuration - object - ``{"autoMount":{"enabled":true,"resources":{}},"hostRoot":"/run/cilium/cgroupv2"}`` \* - :spelling:ignore:`cgroup.autoMount.enabled` - Enable auto mount of cgroup2 filesystem. When ``autoMount`` is enabled, cgroup2 filesystem is mounted at ``cgroup.hostRoot`` path on the underlying host and inside the cilium agent pod. If users disable ``autoMount``\ , it's expected that users have mounted cgroup2 filesystem at the specified ``cgroup.hostRoot`` volume, and then the volume will be mounted inside the cilium agent pod at the same path. - bool - ``true`` \* - :spelling:ignore:`cgroup.autoMount.resources` - Init Container Cgroup Automount resource limits & requests - object - ``{}`` \* - :spelling:ignore:`cgroup.hostRoot` - Configure cgroup root where cgroup2 filesystem is mounted on the host (see also: ``cgroup.autoMount``\ ) - string - ``"/run/cilium/cgroupv2"`` \* - :spelling:ignore:`ciliumEndpointSlice` - CiliumEndpointSlice configuration options. - object - ``{"enabled":false,"rateLimits":[{"burst":20,"limit":10,"nodes":0},{"burst":100,"limit":50,"nodes":100}]}`` \* - :spelling:ignore:`ciliumEndpointSlice.enabled` - Enable Cilium EndpointSlice feature. - bool - ``false`` \* - :spelling:ignore:`ciliumEndpointSlice.rateLimits` - List of rate limit options to be used for the CiliumEndpointSlice controller. Each object in the list must have the following fields: nodes: Count of nodes at which to apply the rate limit. limit: The sustained request rate in requests per second. The maximum rate that can be configured is 50. burst: The burst request rate in requests per second. The maximum burst that can be configured is 100. - list - ``[{"burst":20,"limit":10,"nodes":0},{"burst":100,"limit":50,"nodes":100}]`` \* - :spelling:ignore:`cleanBpfState` - Clean all eBPF datapath state from the initContainer of the cilium-agent DaemonSet. WARNING: Use with care! - bool - ``false`` \* - :spelling:ignore:`cleanState` - Clean all local Cilium state from the initContainer of the cilium-agent DaemonSet. Implies cleanBpfState: true. WARNING: Use with care! - bool - ``false`` \* - :spelling:ignore:`cluster.id` - Unique ID of the cluster. Must be unique across all connected clusters and in the range of 1 to 255. Only required for Cluster Mesh, may be 0 if Cluster Mesh is not used. - int - ``0`` \* - :spelling:ignore:`cluster.name` - Name of the cluster. Only required for Cluster Mesh and mutual authentication with SPIRE. It must respect the following constraints: \* It must contain at most 32 characters; \* It must begin and end with a lower case alphanumeric character; \* It may contain lower case alphanumeric characters and dashes between. The "default" name cannot be used if the Cluster ID is different from 0. - string - ``"default"`` \* - :spelling:ignore:`clustermesh.annotations` - Annotations to be added to all top-level clustermesh objects (resources under templates/clustermesh-apiserver and templates/clustermesh-config) - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.affinity` - Affinity for clustermesh.apiserver - object - ``{"podAntiAffinity":{"preferredDuringSchedulingIgnoredDuringExecution":[{"podAffinityTerm":{"labelSelector":{"matchLabels":{"k8s-app":"clustermesh-apiserver"}},"topologyKey":"kubernetes.io/hostname"},"weight":100}]}}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.init.extraArgs` - Additional arguments to ``clustermesh-apiserver etcdinit``. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.init.extraEnv` - Additional environment variables to ``clustermesh-apiserver etcdinit``. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.init.resources` - Specifies the resources for etcd init container in the apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.lifecycle` - lifecycle setting for the etcd container - | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.009343358688056469,
0.06058864668011665,
0.036657653748989105,
0.03680575639009476,
-0.00547952251508832,
0.022040871903300285,
-0.04465797170996666,
-0.054049454629421234,
0.032992079854011536,
0.01089517679065466,
0.010022209025919437,
-0.11750693619251251,
0.03676086664199829,
-0.0016... | 0.043078 |
to ``clustermesh-apiserver etcdinit``. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.init.extraEnv` - Additional environment variables to ``clustermesh-apiserver etcdinit``. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.init.resources` - Specifies the resources for etcd init container in the apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.lifecycle` - lifecycle setting for the etcd container - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.resources` - Specifies the resources for etcd container in the apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.securityContext` - Security context to be added to clustermesh-apiserver etcd containers - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}}`` \* - :spelling:ignore:`clustermesh.apiserver.etcd.storageMedium` - Specifies whether etcd data is stored in a temporary volume backed by the node's default medium, such as disk, SSD or network storage (Disk), or RAM (Memory). The Memory option enables improved etcd read and write performance at the cost of additional memory usage, which counts against the memory limits of the container. - string - ``"Disk"`` \* - :spelling:ignore:`clustermesh.apiserver.extraArgs` - Additional clustermesh-apiserver arguments. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.extraEnv` - Additional clustermesh-apiserver environment variables. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.extraVolumeMounts` - Additional clustermesh-apiserver volumeMounts. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.extraVolumes` - Additional clustermesh-apiserver volumes. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.healthPort` - TCP port for the clustermesh-apiserver health API. - int - ``9880`` \* - :spelling:ignore:`clustermesh.apiserver.image` - Clustermesh API server image. - object - ``{"digest":"","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/clustermesh-apiserver-ci","tag":"latest","useDigest":false}`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.enabled` - Enable KVStoreMesh. KVStoreMesh caches the information retrieved from the remote clusters in the local etcd instance (deprecated - KVStoreMesh will always be enabled once the option is removed). - bool - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.extraArgs` - Additional KVStoreMesh arguments. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.extraEnv` - Additional KVStoreMesh environment variables. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.extraVolumeMounts` - Additional KVStoreMesh volumeMounts. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.healthPort` - TCP port for the KVStoreMesh health API. - int - ``9881`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.kvstoreMode` - Specify the KVStore mode when running KVStoreMesh Supported values: - "internal": remote cluster identities are cached in etcd that runs as a sidecar within ``clustermesh-apiserver`` pod. - "external": ``clustermesh-apiserver`` will sync remote cluster information to the etcd used as kvstore. This can't be enabled with crd identity allocation mode. - string - ``"internal"`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.lifecycle` - lifecycle setting for the KVStoreMesh container - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.readinessProbe` - Configuration for the KVStoreMesh readiness probe. - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.resources` - Resource requests and limits for the KVStoreMesh container - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.kvstoremesh.securityContext` - KVStoreMesh Security context - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}}`` \* - :spelling:ignore:`clustermesh.apiserver.lifecycle` - lifecycle setting for the apiserver container - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.enabled` - Enables exporting apiserver metrics in OpenMetrics format. - bool - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.etcd.enabled` - Enables exporting etcd metrics in OpenMetrics format. - bool - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.etcd.mode` - Set level of detail for etcd metrics; specify 'extensive' to include server side gRPC histogram metrics. - string - ``"basic"`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.etcd.port` - Configure the port the etcd metric server listens on. - int - ``9963`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.kvstoremesh.enabled` - Enables exporting KVStoreMesh metrics in OpenMetrics format. - bool - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.kvstoremesh.port` - Configure the port the KVStoreMesh metric server listens on. - int - ``9964`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.port` - Configure the port the apiserver metric server listens on. - int - ``9962`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.annotations` - Annotations to add to ServiceMonitor clustermesh-apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.enabled` - Enable service monitor. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.etcd.interval` - Interval for scrape metrics (etcd metrics) - string - ``"10s"`` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.005678430665284395,
0.04000389948487282,
0.01118407491594553,
-0.022790303453803062,
0.0628623515367508,
0.04695049673318863,
-0.03832567110657692,
-0.04096214100718498,
-0.03307541459798813,
0.003784340573474765,
0.028513558208942413,
-0.054435670375823975,
0.02849603444337845,
0.03811... | 0.110235 |
``9962`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.annotations` - Annotations to add to ServiceMonitor clustermesh-apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.enabled` - Enable service monitor. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.etcd.interval` - Interval for scrape metrics (etcd metrics) - string - ``"10s"`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.etcd.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor clustermesh-apiserver (etcd metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.etcd.relabelings` - Relabeling configs for the ServiceMonitor clustermesh-apiserver (etcd metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.etcd.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.interval` - Interval for scrape metrics (apiserver metrics) - string - ``"10s"`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.kvstoremesh.interval` - Interval for scrape metrics (KVStoreMesh metrics) - string - ``"10s"`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.kvstoremesh.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor clustermesh-apiserver (KVStoreMesh metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.kvstoremesh.relabelings` - Relabeling configs for the ServiceMonitor clustermesh-apiserver (KVStoreMesh metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.kvstoremesh.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.labels` - Labels to add to ServiceMonitor clustermesh-apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor clustermesh-apiserver (apiserver metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor clustermesh-apiserver (apiserver metrics) - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.metrics.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.nodeSelector` - Node labels for pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`clustermesh.apiserver.podAnnotations` - Annotations to be added to clustermesh-apiserver pods - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`clustermesh.apiserver.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`clustermesh.apiserver.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by ``maxUnavailable: null`` - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.podDisruptionBudget.unhealthyPodEvictionPolicy` - How are unhealthy, but running, pods counted for eviction - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.podLabels` - Labels to be added to clustermesh-apiserver pods - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.podSecurityContext` - Security context to be added to clustermesh-apiserver pods - object - ``{"fsGroup":65532,"runAsGroup":65532,"runAsNonRoot":true,"runAsUser":65532}`` \* - :spelling:ignore:`clustermesh.apiserver.priorityClassName` - The priority class to use for clustermesh-apiserver - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.readinessProbe` - Configuration for the clustermesh-apiserver readiness probe. - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.replicas` - Number of replicas run for the clustermesh-apiserver deployment. - int - ``1`` \* - :spelling:ignore:`clustermesh.apiserver.resources` - Resource requests and limits for the clustermesh-apiserver - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.securityContext` - Security context to be added to clustermesh-apiserver containers - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}}`` \* - :spelling:ignore:`clustermesh.apiserver.service.annotations` - Annotations for the clustermesh-apiserver service. Example annotations to configure an internal load balancer on different cloud providers: \* AKS: service.beta.kubernetes.io/azure-load-balancer-internal: "true" \* EKS: service.beta.kubernetes.io/aws-load-balancer-scheme: "internal" \* GKE: networking.gke.io/load-balancer-type: "Internal" - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.service.enableSessionAffinity` - Defines when to enable session affinity. Each replica in a clustermesh-apiserver deployment runs its own discrete etcd cluster. Remote clients connect to one of the replicas through a shared Kubernetes Service. A client reconnecting to a different backend will require a full resync to ensure data integrity. Session affinity can reduce the likelihood of this happening, but may not be supported by all cloud providers. Possible values: - "HAOnly" (default) Only enable session affinity for deployments with more than 1 replica. - "Always" Always enable session affinity. - "Never" Never enable session affinity. Useful in environments where session affinity is not supported, but may lead to slightly degraded | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.0710064172744751,
0.0009141892078332603,
-0.01989896409213543,
0.0011635335395112634,
0.01998465321958065,
-0.04296526312828064,
-0.012097105383872986,
-0.044128406792879105,
0.008863252587616444,
0.01783037930727005,
0.04009163752198219,
-0.09116016328334808,
0.030701352283358574,
0.03... | 0.181639 |
not be supported by all cloud providers. Possible values: - "HAOnly" (default) Only enable session affinity for deployments with more than 1 replica. - "Always" Always enable session affinity. - "Never" Never enable session affinity. Useful in environments where session affinity is not supported, but may lead to slightly degraded performance due to more frequent reconnections. - string - ``"HAOnly"`` \* - :spelling:ignore:`clustermesh.apiserver.service.externalTrafficPolicy` - The externalTrafficPolicy of service used for apiserver access. - string - ``"Cluster"`` \* - :spelling:ignore:`clustermesh.apiserver.service.externallyCreated` - Set externallyCreated to true to create the clustermesh-apiserver service outside this helm chart. For example after external load balancer controllers are created. - bool - ``false`` \* - :spelling:ignore:`clustermesh.apiserver.service.internalTrafficPolicy` - The internalTrafficPolicy of service used for apiserver access. - string - ``"Cluster"`` \* - :spelling:ignore:`clustermesh.apiserver.service.labels` - Labels for the clustermesh-apiserver service. - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.service.loadBalancerClass` - Configure a loadBalancerClass. Allows to configure the loadBalancerClass on the clustermesh-apiserver LB service in case the Service type is set to LoadBalancer (requires Kubernetes 1.24+). - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.service.loadBalancerIP` - Configure a specific loadBalancerIP. Allows to configure a specific loadBalancerIP on the clustermesh-apiserver LB service in case the Service type is set to LoadBalancer. - string - ``nil`` \* - :spelling:ignore:`clustermesh.apiserver.service.loadBalancerSourceRanges` - Configure loadBalancerSourceRanges. Allows to configure the source IP ranges allowed to access the clustermesh-apiserver LB service in case the Service type is set to LoadBalancer. - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.service.nodePort` - Optional port to use as the node port for apiserver access. - int - ``32379`` \* - :spelling:ignore:`clustermesh.apiserver.service.type` - The type of service used for apiserver access. - string - ``"NodePort"`` \* - :spelling:ignore:`clustermesh.apiserver.terminationGracePeriodSeconds` - terminationGracePeriodSeconds for the clustermesh-apiserver deployment - int - ``30`` \* - :spelling:ignore:`clustermesh.apiserver.tls.admin` - base64 encoded PEM values for the clustermesh-apiserver admin certificate and private key. Used if 'auto' is not enabled. - object - ``{"cert":"","key":""}`` \* - :spelling:ignore:`clustermesh.apiserver.tls.admin.cert` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tls.admin.key` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tls.authMode` - Configure the clustermesh authentication mode. Supported values: - legacy: All clusters access remote clustermesh instances with the same username (i.e., remote). The "remote" certificate must be generated with CN=remote if provided manually. - migration: Intermediate mode required to upgrade from legacy to cluster (and vice versa) with no disruption. Specifically, it enables the creation of the per-cluster usernames, while still using the common one for authentication. The "remote" certificate must be generated with CN=remote if provided manually (same as legacy). - cluster: Each cluster accesses remote etcd instances with a username depending on the local cluster name (i.e., remote-\ :raw-html-m2r:``\ ). The "remote" certificate must be generated with CN=remote-\ :raw-html-m2r:`` if provided manually. Cluster mode is meaningful only when the same CA is shared across all clusters part of the mesh. - string - ``"migration"`` \* - :spelling:ignore:`clustermesh.apiserver.tls.auto` - Configure automatic TLS certificates generation. A Kubernetes CronJob is used the generate any certificates not provided by the user at installation time. - object - ``{"certManagerIssuerRef":{},"certValidityDuration":1095,"enabled":true,"method":"helm"}`` \* - :spelling:ignore:`clustermesh.apiserver.tls.auto.certManagerIssuerRef` - certmanager issuer used when clustermesh.apiserver.tls.auto.method=certmanager. - object - ``{}`` \* - :spelling:ignore:`clustermesh.apiserver.tls.auto.certValidityDuration` - Generated certificates validity duration in days. - int - ``1095`` \* - :spelling:ignore:`clustermesh.apiserver.tls.auto.enabled` - When set to true, automatically generate a CA and certificates to enable mTLS between clustermesh-apiserver and external workload instances. When set to false you need to pre-create the following secrets: - clustermesh-apiserver-server-cert - clustermesh-apiserver-admin-cert - clustermesh-apiserver-remote-cert - clustermesh-apiserver-local-cert The above secret should at least contains the keys ``tls.crt`` and | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.0006740277749486268,
-0.008537212386727333,
-0.029238060116767883,
-0.00255289813503623,
0.021583303809165955,
-0.0008622645982541144,
-0.04087388515472412,
0.0010962968226522207,
0.016880370676517487,
0.05159018933773041,
0.016654523089528084,
-0.030232861638069153,
0.04381190240383148,
... | 0.080807 |
- When set to true, automatically generate a CA and certificates to enable mTLS between clustermesh-apiserver and external workload instances. When set to false you need to pre-create the following secrets: - clustermesh-apiserver-server-cert - clustermesh-apiserver-admin-cert - clustermesh-apiserver-remote-cert - clustermesh-apiserver-local-cert The above secret should at least contains the keys ``tls.crt`` and ``tls.key`` and optionally ``ca.crt`` if a CA bundle is not configured. - bool - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.tls.enableSecrets` - Allow users to provide their own certificates Users may need to provide their certificates using a mechanism that requires they provide their own secrets. This setting does not apply to any of the auto-generated mechanisms below, it only restricts the creation of secrets via the ``tls-provided`` templates. This option is deprecated as secrets are expected to be created externally when 'auto' is not enabled. - deprecated - ``true`` \* - :spelling:ignore:`clustermesh.apiserver.tls.remote` - base64 encoded PEM values for the clustermesh-apiserver remote cluster certificate and private key. Used if 'auto' is not enabled. - object - ``{"cert":"","key":""}`` \* - :spelling:ignore:`clustermesh.apiserver.tls.remote.cert` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tls.remote.key` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tls.server` - base64 encoded PEM values for the clustermesh-apiserver server certificate and private key. Used if 'auto' is not enabled. - object - ``{"cert":"","extraDnsNames":[],"extraIpAddresses":[],"key":""}`` \* - :spelling:ignore:`clustermesh.apiserver.tls.server.cert` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tls.server.extraDnsNames` - Extra DNS names added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.tls.server.extraIpAddresses` - Extra IP addresses added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.tls.server.key` - Deprecated, as secrets will always need to be created externally if ``auto`` is disabled. - string - ``""`` \* - :spelling:ignore:`clustermesh.apiserver.tolerations` - Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.topologySpreadConstraints` - Pod topology spread constraints for clustermesh-apiserver - list - ``[]`` \* - :spelling:ignore:`clustermesh.apiserver.updateStrategy` - clustermesh-apiserver update strategy - object - ``{"rollingUpdate":{"maxSurge":1,"maxUnavailable":0},"type":"RollingUpdate"}`` \* - :spelling:ignore:`clustermesh.cacheTTL` - The time to live for the cache of a remote cluster after connectivity is lost. If the connection is not re-established within this duration, the cached data is revoked to prevent stale state. If not specified or set to 0s, the cache is never revoked (default). - string - ``"0s"`` \* - :spelling:ignore:`clustermesh.config` - Clustermesh explicit configuration. - object - ``{"clusters":[],"domain":"mesh.cilium.io","enabled":false}`` \* - :spelling:ignore:`clustermesh.config.clusters` - Clusters to be peered in the mesh. @schema type: [object, array] @schema - list - ``[]`` \* - :spelling:ignore:`clustermesh.config.domain` - Default dns domain for the Clustermesh API servers This is used in the case cluster addresses are not provided and IPs are used. - string - ``"mesh.cilium.io"`` \* - :spelling:ignore:`clustermesh.config.enabled` - Enable the Clustermesh explicit configuration. If set to false, you need to provide the following resources yourself: - (Secret) cilium-clustermesh (used by cilium-agent/cilium-operator to connect to the local etcd instance if KVStoreMesh is enabled or the remote clusters if KVStoreMesh is disabled) - (Secret) cilium-kvstoremesh (used by KVStoreMesh to connect to the remote clusters) - (ConfigMap) clustermesh-remote-users (used to create one etcd user per remote cluster if clustermesh-apiserver is used and ``clustermesh.apiserver.tls.authMode`` is not set to ``legacy``\ ) - bool - ``false`` \* - :spelling:ignore:`clustermesh.enableEndpointSliceSynchronization` - Enable the synchronization of Kubernetes EndpointSlices corresponding to the remote endpoints of appropriately-annotated global services through ClusterMesh - bool - ``false`` \* - :spelling:ignore:`clustermesh.enableMCSAPISupport` - Enable Multi-Cluster Services API support (deprecated; use clustermesh.mcsapi.enabled) - bool - | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.003443072782829404,
0.014620027504861355,
0.019382426515221596,
0.060634028166532516,
-0.0065552773885428905,
-0.011482012458145618,
-0.06058194115757942,
-0.01614765077829361,
0.04678697884082794,
0.010760924778878689,
0.031585726886987686,
-0.11093364655971527,
0.10538634657859802,
0.... | 0.054986 |
is not set to ``legacy``\ ) - bool - ``false`` \* - :spelling:ignore:`clustermesh.enableEndpointSliceSynchronization` - Enable the synchronization of Kubernetes EndpointSlices corresponding to the remote endpoints of appropriately-annotated global services through ClusterMesh - bool - ``false`` \* - :spelling:ignore:`clustermesh.enableMCSAPISupport` - Enable Multi-Cluster Services API support (deprecated; use clustermesh.mcsapi.enabled) - bool - ``false`` \* - :spelling:ignore:`clustermesh.maxConnectedClusters` - The maximum number of clusters to support in a ClusterMesh. This value cannot be changed on running clusters, and all clusters in a ClusterMesh must be configured with the same value. Values > 255 will decrease the maximum allocatable cluster-local identities. Supported values are 255 and 511. - int - ``255`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.affinity` - Affinity for coredns-mcsapi-autoconfig - object - ``{}`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.annotations` - Annotations to be added to the coredns-mcsapi-autoconfig Job - object - ``{}`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.clusterDomain` - The cluster domain for the cluster CoreDNS service - string - ``"cluster.local"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.clustersetDomain` - The clusterset domain for the cluster CoreDNS service - string - ``"clusterset.local"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.configMapName` - The ConfigMap name for the cluster CoreDNS service - string - ``"coredns"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.deploymentName` - The Deployment for the cluster CoreDNS service - string - ``"coredns"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.namespace` - The namespace for the cluster CoreDNS service - string - ``"kube-system"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.coredns.serviceAccountName` - The Service Account name for the cluster CoreDNS service - string - ``"coredns"`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.enabled` - Enable auto-configuration of CoreDNS for Multi-Cluster Services API. CoreDNS MUST be at least in version v1.12.2 to run this. - bool - ``false`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.extraArgs` - Additional arguments to ``clustermesh-apiserver coredns-mcsapi-auto-configure``. - list - ``[]`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.extraVolumeMounts` - Additional coredns-mcsapi-autoconfig volumeMounts. - list - ``[]`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.extraVolumes` - Additional coredns-mcsapi-autoconfig volumes. - list - ``[]`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.nodeSelector` - Node selector for coredns-mcsapi-autoconfig ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{}`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.podLabels` - Labels to be added to coredns-mcsapi-autoconfig pods - object - ``{}`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.priorityClassName` - Priority class for coredns-mcsapi-autoconfig ref: https://kubernetes.io/docs/concepts/scheduling-eviction/pod-priority-preemption/#priorityclass - string - ``""`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.resources` - Resource limits for coredns-mcsapi-autoconfig ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers - object - ``{}`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.tolerations` - Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`clustermesh.mcsapi.corednsAutoConfigure.ttlSecondsAfterFinished` - Seconds after which the completed job pod will be deleted - int - ``1800`` \* - :spelling:ignore:`clustermesh.mcsapi.enabled` - Enable Multi-Cluster Services API support - bool - ``false`` \* - :spelling:ignore:`clustermesh.mcsapi.installCRDs` - Enabled MCS-API CRDs auto-installation - bool - ``true`` \* - :spelling:ignore:`clustermesh.policyDefaultLocalCluster` - Control whether policy rules assume by default the local cluster if not explicitly selected - bool - ``true`` \* - :spelling:ignore:`clustermesh.useAPIServer` - Deploy clustermesh-apiserver for clustermesh. This option is typically used with ``clustermesh.config.enabled=true``. Refer to the ``clustermesh.config.enabled=true``\ documentation for more information. - bool - ``false`` \* - :spelling:ignore:`cni.binPath` - Configure the path to the CNI binary directory on the host. - string - ``"/opt/cni/bin"`` \* - :spelling:ignore:`cni.chainingMode` - Configure chaining on top of other CNI plugins. Possible values: - none - aws-cni - flannel - generic-veth - portmap - string - ``nil`` \* - :spelling:ignore:`cni.chainingTarget` - A CNI network name in to which the Cilium plugin should be added as a chained plugin. This will cause the agent to watch for a CNI network with this network name. When it is found, this will be used as the basis for Cilium's CNI configuration file. If this is set, it assumes a chaining mode of generic-veth. As a special case, a chaining mode of aws-cni implies a chainingTarget of aws-cni. - string - ``nil`` \* - :spelling:ignore:`cni.confFileMountPath` - Configure the path to where to mount the ConfigMap inside | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.04769212007522583,
-0.058505602180957794,
0.014537530951201916,
-0.034858886152505875,
0.01393676083534956,
0.020973054692149162,
-0.04868389666080475,
-0.05246348679065704,
0.013627281412482262,
0.04812018945813179,
0.022153638303279877,
-0.07153329253196716,
0.02825724519789219,
0.0374... | 0.156934 |
the basis for Cilium's CNI configuration file. If this is set, it assumes a chaining mode of generic-veth. As a special case, a chaining mode of aws-cni implies a chainingTarget of aws-cni. - string - ``nil`` \* - :spelling:ignore:`cni.confFileMountPath` - Configure the path to where to mount the ConfigMap inside the agent pod. - string - ``"/tmp/cni-configuration"`` \* - :spelling:ignore:`cni.confPath` - Configure the path to the CNI configuration directory on the host. - string - ``"/etc/cni/net.d"`` \* - :spelling:ignore:`cni.configMap` - When defined, configMap will mount the provided value as ConfigMap and interpret the 'cni.configMapKey' value as CNI configuration file and write it when the agent starts up. - string - ``""`` \* - :spelling:ignore:`cni.configMapKey` - Configure the key in the CNI ConfigMap to read the contents of the CNI configuration from. For this to be effective, the 'cni.configMap' parameter must be specified too. Note that the 'cni.configMap' parameter is the name of the ConfigMap, while 'cni.configMapKey' is the name of the key in the ConfigMap data containing the actual configuration. - string - ``"cni-config"`` \* - :spelling:ignore:`cni.customConf` - Skip writing of the CNI configuration. This can be used if writing of the CNI configuration is performed by external automation. - bool - ``false`` \* - :spelling:ignore:`cni.enableRouteMTUForCNIChaining` - Enable route MTU for pod netns when CNI chaining is used - bool - ``false`` \* - :spelling:ignore:`cni.exclusive` - Make Cilium take ownership over the ``/etc/cni/net.d`` directory on the node, renaming all non-Cilium CNI configurations to ``\*.cilium\_bak``. This ensures no Pods can be scheduled using other CNI plugins during Cilium agent downtime. - bool - ``true`` \* - :spelling:ignore:`cni.hostConfDirMountPath` - Configure the path to where the CNI configuration directory is mounted inside the agent pod. - string - ``"/host/etc/cni/net.d"`` \* - :spelling:ignore:`cni.install` - Install the CNI configuration and binary files into the filesystem. - bool - ``true`` \* - :spelling:ignore:`cni.iptablesRemoveAWSRules` - Enable the removal of iptables rules created by the AWS CNI VPC plugin. - bool - ``true`` \* - :spelling:ignore:`cni.logFile` - Configure the log file for CNI logging with retention policy of 7 days. Disable CNI file logging by setting this field to empty explicitly. - string - ``"/var/run/cilium/cilium-cni.log"`` \* - :spelling:ignore:`cni.resources` - Specifies the resources for the cni initContainer - object - ``{"limits":{"cpu":1,"memory":"1Gi"},"requests":{"cpu":"100m","memory":"10Mi"}}`` \* - :spelling:ignore:`cni.uninstall` - Remove the CNI configuration and binary files on agent shutdown. Enable this if you're removing Cilium from the cluster. Disable this to prevent the CNI configuration file from being removed during agent upgrade, which can cause nodes to go unmanageable. - bool - ``false`` \* - :spelling:ignore:`commonLabels` - commonLabels allows users to add common labels for all Cilium resources. - object - ``{}`` \* - :spelling:ignore:`connectivityProbeFrequencyRatio` - Ratio of the connectivity probe frequency vs resource usage, a float in [0, 1]. 0 will give more frequent probing, 1 will give less frequent probing. Probing frequency is dynamically adjusted based on the cluster size. - float64 - ``0.5`` \* - :spelling:ignore:`conntrackGCInterval` - Configure how frequently garbage collection should occur for the datapath connection tracking table. - string - ``"0s"`` \* - :spelling:ignore:`conntrackGCMaxInterval` - Configure the maximum frequency for the garbage collection of the connection tracking table. Only affects the automatic computation for the frequency and has no effect when 'conntrackGCInterval' is set. This can be set to more frequently clean up unused identities created from ToFQDN policies. - string - ``""`` \* - :spelling:ignore:`crdWaitTimeout` - Configure timeout in which Cilium will exit if CRDs are not available - string - ``"5m"`` \* - :spelling:ignore:`daemon.allowedConfigOverrides` - allowedConfigOverrides is a list of config-map keys that can be overridden. That is to say, if this value is set, | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.007622543256729841,
0.012769537977874279,
-0.06906981766223907,
-0.05675726383924484,
-0.00874241441488266,
-0.06273677945137024,
-0.01665841042995453,
0.020888838917016983,
0.01113656535744667,
0.00812483485788107,
0.027752067893743515,
-0.10157255828380585,
0.07078129798173904,
-0.0272... | 0.13189 |
ToFQDN policies. - string - ``""`` \* - :spelling:ignore:`crdWaitTimeout` - Configure timeout in which Cilium will exit if CRDs are not available - string - ``"5m"`` \* - :spelling:ignore:`daemon.allowedConfigOverrides` - allowedConfigOverrides is a list of config-map keys that can be overridden. That is to say, if this value is set, config sources (excepting the first one) can only override keys in this list. This takes precedence over blockedConfigOverrides. By default, all keys may be overridden. To disable overrides, set this to "none" or change the configSources variable. - string - ``nil`` \* - :spelling:ignore:`daemon.blockedConfigOverrides` - blockedConfigOverrides is a list of config-map keys that may not be overridden. In other words, if any of these keys appear in a configuration source excepting the first one, they will be ignored This is ignored if allowedConfigOverrides is set. By default, all keys may be overridden. - string - ``nil`` \* - :spelling:ignore:`daemon.configSources` - Configure a custom list of possible configuration override sources The default is "config-map:cilium-config,cilium-node-config". For supported values, see the help text for the build-config subcommand. Note that this value should be a comma-separated string. - string - ``nil`` \* - :spelling:ignore:`daemon.enableSourceIPVerification` - enableSourceIPVerification is a boolean flag to enable or disable the Source IP verification of endpoints. This flag is useful when Cilium is chained with other CNIs. By default, this functionality is enabled - bool - ``true`` \* - :spelling:ignore:`daemon.runPath` - Configure where Cilium runtime state should be stored. - string - ``"/var/run/cilium"`` \* - :spelling:ignore:`dashboards` - Grafana dashboards for cilium-agent grafana can import dashboards based on the label and value ref: https://github.com/grafana/helm-charts/tree/main/charts/grafana#sidecar-for-dashboards - object - ``{"annotations":{},"enabled":false,"label":"grafana\_dashboard","labelValue":"1","namespace":null}`` \* - :spelling:ignore:`debug.enabled` - Enable debug logging - bool - ``false`` \* - :spelling:ignore:`debug.metricsSamplingInterval` - Set the agent-internal metrics sampling frequency. This sets the frequency of the internal sampling of the agent metrics. These are available via the "cilium-dbg shell -- metrics -s" command and are part of the metrics HTML page included in the sysdump. @schema type: [null, string] @schema - string - ``"5m"`` \* - :spelling:ignore:`debug.verbose` - Configure verbosity levels for debug logging This option is used to enable debug messages for operations related to such sub-system such as (e.g. kvstore, envoy, datapath, policy, or tagged), and flow is for enabling debug messages emitted per request, message and connection. Multiple values can be set via a space-separated string (e.g. "datapath envoy"). Applicable values: - flow - kvstore - envoy - datapath - policy - tagged - string - ``nil`` \* - :spelling:ignore:`defaultLBServiceIPAM` - defaultLBServiceIPAM indicates the default LoadBalancer Service IPAM when no LoadBalancer class is set. Applicable values: lbipam, nodeipam, none - string - ``"lbipam"`` \* - :spelling:ignore:`directRoutingSkipUnreachable` - Enable skipping of PodCIDR routes between worker nodes if the worker nodes are in a different L2 network segment. - bool - ``false`` \* - :spelling:ignore:`disableEndpointCRD` - Disable the usage of CiliumEndpoint CRD. - bool - ``false`` \* - :spelling:ignore:`dnsPolicy` - DNS policy for Cilium agent pods. Ref: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy - string - ``""`` \* - :spelling:ignore:`dnsProxy.dnsRejectResponseCode` - DNS response code for rejecting DNS requests, available options are '[nameError refused]'. - string - ``"refused"`` \* - :spelling:ignore:`dnsProxy.enableDnsCompression` - Allow the DNS proxy to compress responses to endpoints that are larger than 512 Bytes or the EDNS0 option, if present. - bool - ``true`` \* - :spelling:ignore:`dnsProxy.endpointMaxIpPerHostname` - Maximum number of IPs to maintain per FQDN name for each endpoint. - int - ``1000`` \* - :spelling:ignore:`dnsProxy.idleConnectionGracePeriod` - Time during which idle but previously active connections with expired DNS lookups are still considered alive. - string - ``"0s"`` \* - :spelling:ignore:`dnsProxy.maxDeferredConnectionDeletes` - Maximum number of IPs to retain for expired DNS lookups with still-active | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.058121535927057266,
0.003774728626012802,
-0.08538056164979935,
-0.01574673503637314,
-0.0070197684690356255,
-0.00003071271567023359,
0.019237210974097252,
-0.02084284834563732,
-0.015528310090303421,
0.002347925677895546,
0.0693352222442627,
-0.01646965928375721,
0.03467920050024986,
... | 0.045066 |
per FQDN name for each endpoint. - int - ``1000`` \* - :spelling:ignore:`dnsProxy.idleConnectionGracePeriod` - Time during which idle but previously active connections with expired DNS lookups are still considered alive. - string - ``"0s"`` \* - :spelling:ignore:`dnsProxy.maxDeferredConnectionDeletes` - Maximum number of IPs to retain for expired DNS lookups with still-active connections. - int - ``10000`` \* - :spelling:ignore:`dnsProxy.minTtl` - The minimum time, in seconds, to use DNS data for toFQDNs policies. If the upstream DNS server returns a DNS record with a shorter TTL, Cilium overwrites the TTL with this value. Setting this value to zero means that Cilium will honor the TTLs returned by the upstream DNS server. - int - ``0`` \* - :spelling:ignore:`dnsProxy.preAllocateIdentities` - Pre-allocate ToFQDN identities. This reduces DNS proxy tail latency, at the potential cost of some unnecessary policymap entries. Disable this if you have a large (200+) number of unique ToFQDN selectors. - bool - ``true`` \* - :spelling:ignore:`dnsProxy.preCache` - DNS cache data at this path is preloaded on agent startup. - string - ``""`` \* - :spelling:ignore:`dnsProxy.proxyPort` - Global port on which the in-agent DNS proxy should listen. Default 0 is a OS-assigned port. - int - ``0`` \* - :spelling:ignore:`dnsProxy.proxyResponseMaxDelay` - The maximum time the DNS proxy holds an allowed DNS response before sending it along. Responses are sent as soon as the datapath is updated with the new IP information. - string - ``"100ms"`` \* - :spelling:ignore:`dnsProxy.socketLingerTimeout` - Timeout (in seconds) when closing the connection between the DNS proxy and the upstream server. If set to 0, the connection is closed immediately (with TCP RST). If set to -1, the connection is closed asynchronously in the background. - int - ``10`` \* - :spelling:ignore:`egressGateway.enabled` - Enables egress gateway to redirect and SNAT the traffic that leaves the cluster. - bool - ``false`` \* - :spelling:ignore:`egressGateway.reconciliationTriggerInterval` - Time between triggers of egress gateway state reconciliations - string - ``"1s"`` \* - :spelling:ignore:`enableCriticalPriorityClass` - Explicitly enable or disable priority class. .Capabilities.KubeVersion is unsettable in ``helm template`` calls, it depends on k8s libraries version that Helm was compiled against. This option allows to explicitly disable setting the priority class, which is useful for rendering charts for gke clusters in advance. - bool - ``true`` \* - :spelling:ignore:`enableIPv4BIGTCP` - Enables IPv4 BIG TCP support which increases maximum IPv4 GSO/GRO limits for nodes and pods - bool - ``false`` \* - :spelling:ignore:`enableIPv4Masquerade` - Enables masquerading of IPv4 traffic leaving the node from endpoints. - bool - ``true`` unless ipam eni mode is active \* - :spelling:ignore:`enableIPv6BIGTCP` - Enables IPv6 BIG TCP support which increases maximum IPv6 GSO/GRO limits for nodes and pods - bool - ``false`` \* - :spelling:ignore:`enableIPv6Masquerade` - Enables masquerading of IPv6 traffic leaving the node from endpoints. - bool - ``true`` \* - :spelling:ignore:`enableInternalTrafficPolicy` - Enable Internal Traffic Policy - bool - ``true`` \* - :spelling:ignore:`enableLBIPAM` - Enable LoadBalancer IP Address Management - bool - ``true`` \* - :spelling:ignore:`enableMasqueradeRouteSource` - Enables masquerading to the source of the route for traffic leaving the node from endpoints. - bool - ``false`` \* - :spelling:ignore:`enableNoServiceEndpointsRoutable` - Enable routing to a service that has zero endpoints - bool - ``true`` \* - :spelling:ignore:`enableNonDefaultDenyPolicies` - Enable Non-Default-Deny policies - bool - ``true`` \* - :spelling:ignore:`enableTunnelBIGTCP` - Enable BIG TCP in tunneling mode and increase maximum GRO/GSO limits for VXLAN/GENEVE tunnels - bool - ``false`` \* - :spelling:ignore:`enableXTSocketFallback` - Enables the fallback compatibility solution for when the xt\_socket kernel module is missing and it is needed for the datapath L7 redirection to work properly. See documentation for details on when this can be disabled: https://docs.cilium.io/en/stable/operations/system\_requirements/#linux-kernel. - bool - ``true`` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.1283930093050003,
0.0035587900783866644,
-0.04828561842441559,
-0.022336916998028755,
-0.07958134263753891,
-0.06796785444021225,
-0.02474582940340042,
-0.010633949190378189,
0.04122929275035858,
-0.03581595420837402,
0.0077147879637777805,
0.047361522912979126,
-0.00045881260302849114,
... | 0.084031 |
VXLAN/GENEVE tunnels - bool - ``false`` \* - :spelling:ignore:`enableXTSocketFallback` - Enables the fallback compatibility solution for when the xt\_socket kernel module is missing and it is needed for the datapath L7 redirection to work properly. See documentation for details on when this can be disabled: https://docs.cilium.io/en/stable/operations/system\_requirements/#linux-kernel. - bool - ``true`` \* - :spelling:ignore:`encryption.enabled` - Enable transparent network encryption. - bool - ``false`` \* - :spelling:ignore:`encryption.ipsec.encryptedOverlay` - Enable IPsec encrypted overlay - bool - ``false`` \* - :spelling:ignore:`encryption.ipsec.interface` - The interface to use for encrypted traffic. - string - ``""`` \* - :spelling:ignore:`encryption.ipsec.keyFile` - Name of the key file inside the Kubernetes secret configured via secretName. - string - ``"keys"`` \* - :spelling:ignore:`encryption.ipsec.keyRotationDuration` - Maximum duration of the IPsec key rotation. The previous key will be removed after that delay. - string - ``"5m"`` \* - :spelling:ignore:`encryption.ipsec.keyWatcher` - Enable the key watcher. If disabled, a restart of the agent will be necessary on key rotations. - bool - ``true`` \* - :spelling:ignore:`encryption.ipsec.mountPath` - Path to mount the secret inside the Cilium pod. - string - ``"/etc/ipsec"`` \* - :spelling:ignore:`encryption.ipsec.secretName` - Name of the Kubernetes secret containing the encryption keys. - string - ``"cilium-ipsec-keys"`` \* - :spelling:ignore:`encryption.nodeEncryption` - Enable encryption for pure node to node traffic. This option is only effective when encryption.type is set to "wireguard". - bool - ``false`` \* - :spelling:ignore:`encryption.strictMode` - Configure the Encryption Pod2Pod strict mode. - object - ``{"allowRemoteNodeIdentities":false,"cidr":"","egress":{"allowRemoteNodeIdentities":false,"cidr":"","enabled":false},"enabled":false,"ingress":{"enabled":false}}`` \* - :spelling:ignore:`encryption.strictMode.allowRemoteNodeIdentities` - Allow dynamic lookup of remote node identities. (deprecated: please use encryption.strictMode.egress.allowRemoteNodeIdentities) This is required when tunneling is used or direct routing is used and the node CIDR and pod CIDR overlap. - bool - ``false`` \* - :spelling:ignore:`encryption.strictMode.cidr` - CIDR for the Encryption Pod2Pod strict mode. (deprecated: please use encryption.strictMode.egress.cidr) - string - ``""`` \* - :spelling:ignore:`encryption.strictMode.egress.allowRemoteNodeIdentities` - Allow dynamic lookup of remote node identities. This is required when tunneling is used or direct routing is used and the node CIDR and pod CIDR overlap. - bool - ``false`` \* - :spelling:ignore:`encryption.strictMode.egress.cidr` - CIDR for the Encryption Pod2Pod strict egress mode. - string - ``""`` \* - :spelling:ignore:`encryption.strictMode.egress.enabled` - Enable strict egress encryption. - bool - ``false`` \* - :spelling:ignore:`encryption.strictMode.enabled` - Enable Encryption Pod2Pod strict mode. (deprecated: please use encryption.strictMode.egress.enabled) - bool - ``false`` \* - :spelling:ignore:`encryption.strictMode.ingress.enabled` - Enable strict ingress encryption. When enabled, all unencrypted overlay ingress traffic will be dropped. This option is only applicable when WireGuard and tunneling are enabled. - bool - ``false`` \* - :spelling:ignore:`encryption.type` - Encryption method. Can be one of ipsec, wireguard or ztunnel. - string - ``"ipsec"`` \* - :spelling:ignore:`encryption.wireguard.persistentKeepalive` - Controls WireGuard PersistentKeepalive option. Set 0s to disable. - string - ``"0s"`` \* - :spelling:ignore:`endpointHealthChecking.enabled` - Enable connectivity health checking between virtual endpoints. - bool - ``true`` \* - :spelling:ignore:`endpointLockdownOnMapOverflow` - Enable endpoint lockdown on policy map overflow. - bool - ``false`` \* - :spelling:ignore:`endpointRoutes.enabled` - Enable use of per endpoint routes instead of routing via the cilium\_host interface. - bool - ``false`` \* - :spelling:ignore:`eni.awsEnablePrefixDelegation` - Enable ENI prefix delegation - bool - ``false`` \* - :spelling:ignore:`eni.awsReleaseExcessIPs` - Release IPs not used from the ENI - bool - ``false`` \* - :spelling:ignore:`eni.ec2APIEndpoint` - EC2 API endpoint to use - string - ``""`` \* - :spelling:ignore:`eni.enabled` - Enable Elastic Network Interface (ENI) integration. - bool - ``false`` \* - :spelling:ignore:`eni.eniTags` - Tags to apply to the newly created ENIs - object - ``{}`` \* - :spelling:ignore:`eni.gcInterval` - Interval for garbage collection of unattached ENIs. Set to "0s" to disable. - string - ``"5m"`` \* - :spelling:ignore:`eni.gcTags` - Additional tags attached to ENIs created by Cilium. Dangling ENIs with this tag will be garbage collected | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.029184293001890182,
-0.006908674258738756,
-0.030635612085461617,
-0.018590468913316727,
0.0028254687786102295,
-0.02427675388753414,
-0.02741367183625698,
0.03138104826211929,
0.021784165874123573,
-0.005513423588126898,
0.02860313653945923,
-0.04555502161383629,
-0.07257189601659775,
... | 0.093154 |
to the newly created ENIs - object - ``{}`` \* - :spelling:ignore:`eni.gcInterval` - Interval for garbage collection of unattached ENIs. Set to "0s" to disable. - string - ``"5m"`` \* - :spelling:ignore:`eni.gcTags` - Additional tags attached to ENIs created by Cilium. Dangling ENIs with this tag will be garbage collected - object - ``{"io.cilium/cilium-managed":"true,"io.cilium/cluster-name":""}`` \* - :spelling:ignore:`eni.iamRole` - If using IAM role for Service Accounts will not try to inject identity values from cilium-aws kubernetes secret. Adds annotation to service account if managed by Helm. See https://github.com/aws/amazon-eks-pod-identity-webhook - string - ``""`` \* - :spelling:ignore:`eni.instanceTagsFilter` - Filter via AWS EC2 Instance tags (k=v) which will dictate which AWS EC2 Instances are going to be used to create new ENIs - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec` - NodeSpec configuration for the ENI - object - ``{"deleteOnTermination":null,"disablePrefixDelegation":false,"excludeInterfaceTags":[],"firstInterfaceIndex":null,"securityGroupTags":[],"securityGroups":[],"subnetIDs":[],"subnetTags":[],"usePrimaryAddress":false}`` \* - :spelling:ignore:`eni.nodeSpec.deleteOnTermination` - Delete ENI on termination @schema type: [null, boolean] @schema - string - ``nil`` \* - :spelling:ignore:`eni.nodeSpec.disablePrefixDelegation` - Disable prefix delegation for IP allocation - bool - ``false`` \* - :spelling:ignore:`eni.nodeSpec.excludeInterfaceTags` - Exclude interface tags to use for IP allocation - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec.firstInterfaceIndex` - First interface index to use for IP allocation @schema type: [null, integer] @schema - string - ``nil`` \* - :spelling:ignore:`eni.nodeSpec.securityGroupTags` - Security group tags to use for IP allocation - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec.securityGroups` - Security groups to use for IP allocation - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec.subnetIDs` - Subnet IDs to use for IP allocation - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec.subnetTags` - Subnet tags to use for IP allocation - list - ``[]`` \* - :spelling:ignore:`eni.nodeSpec.usePrimaryAddress` - Use primary address for IP allocation - bool - ``false`` \* - :spelling:ignore:`eni.subnetIDsFilter` - Filter via subnet IDs which will dictate which subnets are going to be used to create new ENIs Important note: This requires that each instance has an ENI with a matching subnet attached when Cilium is deployed. If you only want to control subnets for ENIs attached by Cilium, use the CNI configuration file settings (cni.customConf) instead. - list - ``[]`` \* - :spelling:ignore:`eni.subnetTagsFilter` - Filter via tags (k=v) which will dictate which subnets are going to be used to create new ENIs Important note: This requires that each instance has an ENI with a matching subnet attached when Cilium is deployed. If you only want to control subnets for ENIs attached by Cilium, use the CNI configuration file settings (cni.customConf) instead. - list - ``[]`` \* - :spelling:ignore:`envoy.affinity` - Affinity for cilium-envoy. - object - ``{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchExpressions":[{"key":"cilium.io/no-schedule","operator":"NotIn","values":["true"]}]}]}},"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]},"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium-envoy"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`envoy.annotations` - Annotations to be added to all top-level cilium-envoy objects (resources under templates/cilium-envoy) - object - ``{}`` \* - :spelling:ignore:`envoy.baseID` - Set Envoy'--base-id' to use when allocating shared memory regions. Only needs to be changed if multiple Envoy instances will run on the same node and may have conflicts. Supported values: 0 - 4294967295. Defaults to '0' - int - ``0`` \* - :spelling:ignore:`envoy.bootstrapConfigMap` - ADVANCED OPTION: Bring your own custom Envoy bootstrap ConfigMap. Provide the name of a ConfigMap with a ``bootstrap-config.json`` key. When specified, Envoy will use this ConfigMap instead of the default provided by the chart. WARNING: Use of this setting has the potential to prevent cilium-envoy from starting up, and can cause unexpected behavior (e.g. due to syntax error or semantically incorrect configuration). Before submitting an issue, please ensure you have disabled this feature, as support cannot be provided for custom Envoy bootstrap configs. @schema type: [null, string] @schema - string - ``nil`` \* - :spelling:ignore:`envoy.clusterMaxConnections` - Maximum number of connections on Envoy clusters - int - ``1024`` \* - :spelling:ignore:`envoy.clusterMaxRequests` - Maximum number | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.005517343524843454,
0.08982086181640625,
-0.034010883420705795,
-0.009000878781080246,
0.02715321257710457,
-0.07935354858636856,
0.052689243108034134,
-0.05556049942970276,
0.08831916749477386,
0.055854327976703644,
0.06603740900754929,
-0.14534282684326172,
0.060703691095113754,
-0.05... | 0.203035 |
submitting an issue, please ensure you have disabled this feature, as support cannot be provided for custom Envoy bootstrap configs. @schema type: [null, string] @schema - string - ``nil`` \* - :spelling:ignore:`envoy.clusterMaxConnections` - Maximum number of connections on Envoy clusters - int - ``1024`` \* - :spelling:ignore:`envoy.clusterMaxRequests` - Maximum number of requests on Envoy clusters - int - ``1024`` \* - :spelling:ignore:`envoy.connectTimeoutSeconds` - Time in seconds after which a TCP connection attempt times out - int - ``2`` \* - :spelling:ignore:`envoy.debug.admin.enabled` - Enable admin interface for cilium-envoy. This is useful for debugging and should not be enabled in production. - bool - ``false`` \* - :spelling:ignore:`envoy.debug.admin.port` - Port number (bound to loopback interface). kubectl port-forward can be used to access the admin interface. - int - ``9901`` \* - :spelling:ignore:`envoy.dnsPolicy` - DNS policy for Cilium envoy pods. Ref: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy - string - ``nil`` \* - :spelling:ignore:`envoy.enabled` - Enable Envoy Proxy in standalone DaemonSet. This field is enabled by default for new installation. - string - ``true`` for new installation \* - :spelling:ignore:`envoy.extraArgs` - Additional envoy container arguments. - list - ``[]`` \* - :spelling:ignore:`envoy.extraContainers` - Additional containers added to the cilium Envoy DaemonSet. - list - ``[]`` \* - :spelling:ignore:`envoy.extraEnv` - Additional envoy container environment variables. - list - ``[]`` \* - :spelling:ignore:`envoy.extraHostPathMounts` - Additional envoy hostPath mounts. - list - ``[]`` \* - :spelling:ignore:`envoy.extraVolumeMounts` - Additional envoy volumeMounts. - list - ``[]`` \* - :spelling:ignore:`envoy.extraVolumes` - Additional envoy volumes. - list - ``[]`` \* - :spelling:ignore:`envoy.healthPort` - TCP port for the health API. - int - ``9878`` \* - :spelling:ignore:`envoy.httpRetryCount` - Maximum number of retries for each HTTP request - int - ``3`` \* - :spelling:ignore:`envoy.httpUpstreamLingerTimeout` - Time in seconds to block Envoy worker thread while an upstream HTTP connection is closing. If set to 0, the connection is closed immediately (with TCP RST). If set to -1, the connection is closed asynchronously in the background. - string - ``nil`` \* - :spelling:ignore:`envoy.idleTimeoutDurationSeconds` - Set Envoy upstream HTTP idle connection timeout seconds. Does not apply to connections with pending requests. Default 60s - int - ``60`` \* - :spelling:ignore:`envoy.image` - Envoy container image. - object - ``{"digest":"sha256:c0d2fbd63c3576d0c397e6c84fe149056e2bbb940d538ab3baebb86f703642d3","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/cilium-envoy","tag":"v1.36.5-1768567475-cb4965cbd69f290aef785c42dfadf6159adca553","useDigest":true}`` \* - :spelling:ignore:`envoy.initContainers` - Init containers added to the cilium Envoy DaemonSet. - list - ``[]`` \* - :spelling:ignore:`envoy.initialFetchTimeoutSeconds` - Time in seconds after which the initial fetch on an xDS stream is considered timed out - int - ``30`` \* - :spelling:ignore:`envoy.livenessProbe.enabled` - Enable liveness probe for cilium-envoy - bool - ``true`` \* - :spelling:ignore:`envoy.livenessProbe.failureThreshold` - failure threshold of liveness probe - int - ``10`` \* - :spelling:ignore:`envoy.livenessProbe.periodSeconds` - interval between checks of the liveness probe - int - ``30`` \* - :spelling:ignore:`envoy.log.accessLogBufferSize` - Size of the Envoy access log buffer created within the agent in bytes. Tune this value up if you encounter "Envoy: Discarded truncated access log message" errors. Large request/response header sizes (e.g. 16KiB) will require a larger buffer size. - int - ``4096`` \* - :spelling:ignore:`envoy.log.defaultLevel` - Default log level of Envoy application log that is configured if Cilium debug / verbose logging isn't enabled. This option allows to have a different log level than the Cilium Agent - e.g. lower it to ``critical``. Possible values: trace, debug, info, warning, error, critical, off - string - Defaults to the default log level of the Cilium Agent - ``info`` \* - :spelling:ignore:`envoy.log.format` - The format string to use for laying out the log message metadata of Envoy. If specified, Envoy will use text format output. This setting is mutually exclusive with envoy.log.format\_json. - string - ``"[%Y-%m-%d %T.%e][%t][%l][%n] [%g:%#] %v"`` \* - :spelling:ignore:`envoy.log.format\_json` - The JSON logging format to | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.02888723835349083,
-0.008044418878853321,
-0.051189225167036057,
0.024682683870196342,
-0.0449538491666317,
-0.03546947240829468,
-0.07254477590322495,
-0.020689325407147408,
0.007332089357078075,
0.06069256737828255,
-0.009970816783607006,
-0.05352728068828583,
-0.01987653411924839,
0.0... | 0.166327 |
``info`` \* - :spelling:ignore:`envoy.log.format` - The format string to use for laying out the log message metadata of Envoy. If specified, Envoy will use text format output. This setting is mutually exclusive with envoy.log.format\_json. - string - ``"[%Y-%m-%d %T.%e][%t][%l][%n] [%g:%#] %v"`` \* - :spelling:ignore:`envoy.log.format\_json` - The JSON logging format to use for Envoy. This setting is mutually exclusive with envoy.log.format. ref: https://www.envoyproxy.io/docs/envoy/latest/api-v3/config/bootstrap/v3/bootstrap.proto#envoy-v3-api-field-config-bootstrap-v3-bootstrap-applicationlogconfig-logformat-json-format - string - ``nil`` \* - :spelling:ignore:`envoy.log.path` - Path to a separate Envoy log file, if any. Defaults to /dev/stdout. - string - ``""`` \* - :spelling:ignore:`envoy.maxConcurrentRetries` - Maximum number of concurrent retries on Envoy clusters - int - ``128`` \* - :spelling:ignore:`envoy.maxConnectionDurationSeconds` - Set Envoy HTTP option max\_connection\_duration seconds. Default 0 (disable) - int - ``0`` \* - :spelling:ignore:`envoy.maxGlobalDownstreamConnections` - Maximum number of global downstream connections - int - ``50000`` \* - :spelling:ignore:`envoy.maxRequestsPerConnection` - ProxyMaxRequestsPerConnection specifies the max\_requests\_per\_connection setting for Envoy - int - ``0`` \* - :spelling:ignore:`envoy.nodeSelector` - Node selector for cilium-envoy. - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`envoy.podAnnotations` - Annotations to be added to envoy pods - object - ``{}`` \* - :spelling:ignore:`envoy.podLabels` - Labels to be added to envoy pods - object - ``{}`` \* - :spelling:ignore:`envoy.podSecurityContext` - Security Context for cilium-envoy pods. - object - ``{"appArmorProfile":{"type":"Unconfined"}}`` \* - :spelling:ignore:`envoy.podSecurityContext.appArmorProfile` - AppArmorProfile options for the ``cilium-agent`` and init containers - object - ``{"type":"Unconfined"}`` \* - :spelling:ignore:`envoy.policyRestoreTimeoutDuration` - Max duration to wait for endpoint policies to be restored on restart. Default "3m". - string - ``nil`` \* - :spelling:ignore:`envoy.priorityClassName` - The priority class to use for cilium-envoy. - string - ``nil`` \* - :spelling:ignore:`envoy.prometheus` - Configure Cilium Envoy Prometheus options. Note that some of these apply to either cilium-agent or cilium-envoy. - object - ``{"enabled":true,"port":"9964","serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","labels":{},"metricRelabelings":null,"relabelings":[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}],"scrapeTimeout":null}}`` \* - :spelling:ignore:`envoy.prometheus.enabled` - Enable prometheus metrics for cilium-envoy - bool - ``true`` \* - :spelling:ignore:`envoy.prometheus.port` - Serve prometheus metrics for cilium-envoy on the configured port - string - ``"9964"`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.annotations` - Annotations to add to ServiceMonitor cilium-envoy - object - ``{}`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.enabled` - Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) Note that this setting applies to both cilium-envoy \*and\* cilium-agent with Envoy enabled. - bool - ``false`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.interval` - Interval for scrape metrics. - string - ``"10s"`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.labels` - Labels to add to ServiceMonitor cilium-envoy - object - ``{}`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor cilium-envoy or for cilium-agent with Envoy configured. - string - ``nil`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor cilium-envoy or for cilium-agent with Envoy configured. - list - ``[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}]`` \* - :spelling:ignore:`envoy.prometheus.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`envoy.readinessProbe.failureThreshold` - failure threshold of readiness probe - int - ``3`` \* - :spelling:ignore:`envoy.readinessProbe.periodSeconds` - interval between checks of the readiness probe - int - ``30`` \* - :spelling:ignore:`envoy.resources` - Envoy resource limits & requests ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ - object - ``{}`` \* - :spelling:ignore:`envoy.rollOutPods` - Roll out cilium envoy pods automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`envoy.securityContext.capabilities.envoy` - Capabilities for the ``cilium-envoy`` container. Even though granted to the container, the cilium-envoy-starter wrapper drops all capabilities after forking the actual Envoy process. ``NET\_BIND\_SERVICE`` is the only capability that can be passed to the Envoy process by setting ``envoy.securityContext.capabilities.keepNetBindService=true`` (in addition to granting the capability to the container). Note: In case of embedded envoy, the capability must be granted to the cilium-agent container. - list - ``["NET\_ADMIN","SYS\_ADMIN"]`` \* - :spelling:ignore:`envoy.securityContext.capabilities.keepCapNetBindService` - Keep capability ``NET\_BIND\_SERVICE`` for Envoy process. - bool - ``false`` \* - :spelling:ignore:`envoy.securityContext.privileged` - Run the pod with elevated privileges - bool | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.018649641424417496,
0.06807749718427658,
-0.005926823243498802,
0.057666949927806854,
0.014480841346085072,
-0.05131983757019043,
0.02306859754025936,
-0.01226194016635418,
0.057857148349285126,
0.007968639954924583,
-0.012220226228237152,
-0.012934339232742786,
-0.01239674724638462,
0.1... | 0.118728 |
the capability to the container). Note: In case of embedded envoy, the capability must be granted to the cilium-agent container. - list - ``["NET\_ADMIN","SYS\_ADMIN"]`` \* - :spelling:ignore:`envoy.securityContext.capabilities.keepCapNetBindService` - Keep capability ``NET\_BIND\_SERVICE`` for Envoy process. - bool - ``false`` \* - :spelling:ignore:`envoy.securityContext.privileged` - Run the pod with elevated privileges - bool - ``false`` \* - :spelling:ignore:`envoy.securityContext.seLinuxOptions` - SELinux options for the ``cilium-envoy`` container - object - ``{"level":"s0","type":"spc\_t"}`` \* - :spelling:ignore:`envoy.startupProbe.enabled` - Enable startup probe for cilium-envoy - bool - ``true`` \* - :spelling:ignore:`envoy.startupProbe.failureThreshold` - failure threshold of startup probe. 105 x 2s translates to the old behaviour of the readiness probe (120s delay + 30 x 3s) - int - ``105`` \* - :spelling:ignore:`envoy.startupProbe.periodSeconds` - interval between checks of the startup probe - int - ``2`` \* - :spelling:ignore:`envoy.streamIdleTimeoutDurationSeconds` - Set Envoy the amount of time that the connection manager will allow a stream to exist with no upstream or downstream activity. default 5 minutes - int - ``300`` \* - :spelling:ignore:`envoy.terminationGracePeriodSeconds` - Configure termination grace period for cilium-envoy DaemonSet. - int - ``1`` \* - :spelling:ignore:`envoy.tolerations` - Node tolerations for envoy scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[{"operator":"Exists"}]`` \* - :spelling:ignore:`envoy.updateStrategy` - cilium-envoy update strategy ref: https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/#updating-a-daemonset - object - ``{"rollingUpdate":{"maxUnavailable":2},"type":"RollingUpdate"}`` \* - :spelling:ignore:`envoy.useOriginalSourceAddress` - For cases when CiliumEnvoyConfig is not used directly (Ingress, Gateway), configures Cilium BPF Metadata listener filter to use the original source address when extracting the metadata for a request. - bool - ``true`` \* - :spelling:ignore:`envoy.xffNumTrustedHopsL7PolicyEgress` - Number of trusted hops regarding the x-forwarded-for and related HTTP headers for the egress L7 policy enforcement Envoy listeners. - int - ``0`` \* - :spelling:ignore:`envoy.xffNumTrustedHopsL7PolicyIngress` - Number of trusted hops regarding the x-forwarded-for and related HTTP headers for the ingress L7 policy enforcement Envoy listeners. - int - ``0`` \* - :spelling:ignore:`envoyConfig.enabled` - Enable CiliumEnvoyConfig CRD CiliumEnvoyConfig CRD can also be implicitly enabled by other options. - bool - ``false`` \* - :spelling:ignore:`envoyConfig.retryInterval` - Interval in which an attempt is made to reconcile failed EnvoyConfigs. If the duration is zero, the retry is deactivated. - string - ``"15s"`` \* - :spelling:ignore:`envoyConfig.secretsNamespace` - SecretsNamespace is the namespace in which envoy SDS will retrieve secrets from. - object - ``{"create":true,"name":"cilium-secrets"}`` \* - :spelling:ignore:`envoyConfig.secretsNamespace.create` - Create secrets namespace for CiliumEnvoyConfig CRDs. - bool - ``true`` \* - :spelling:ignore:`envoyConfig.secretsNamespace.name` - The name of the secret namespace to which Cilium agents are given read access. - string - ``"cilium-secrets"`` \* - :spelling:ignore:`etcd.enabled` - Enable etcd mode for the agent. - bool - ``false`` \* - :spelling:ignore:`etcd.endpoints` - List of etcd endpoints - list - ``["https://CHANGE-ME:2379"]`` \* - :spelling:ignore:`etcd.ssl` - Enable use of TLS/SSL for connectivity to etcd. - bool - ``false`` \* - :spelling:ignore:`extraArgs` - Additional agent container arguments. - list - ``[]`` \* - :spelling:ignore:`extraConfig` - extraConfig allows you to specify additional configuration parameters to be included in the cilium-config configmap. - object - ``{}`` \* - :spelling:ignore:`extraContainers` - Additional containers added to the cilium DaemonSet. - list - ``[]`` \* - :spelling:ignore:`extraEnv` - Additional agent container environment variables. - list - ``[]`` \* - :spelling:ignore:`extraHostPathMounts` - Additional agent hostPath mounts. - list - ``[]`` \* - :spelling:ignore:`extraInitContainers` - Additional initContainers added to the cilium Daemonset. - list - ``[]`` \* - :spelling:ignore:`extraVolumeMounts` - Additional agent volumeMounts. - list - ``[]`` \* - :spelling:ignore:`extraVolumes` - Additional agent volumes. - list - ``[]`` \* - :spelling:ignore:`forceDeviceDetection` - Forces the auto-detection of devices, even if specific devices are explicitly listed - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.enableAlpn` - Enable ALPN for all listeners configured with Gateway API. ALPN will attempt HTTP/2, then HTTP 1.1. Note that this will also | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.040976542979478836,
0.021196968853473663,
-0.06927348673343658,
0.02292253077030182,
0.009281127713620663,
-0.03457366302609444,
0.009365664795041084,
-0.00006574756116606295,
-0.009048137813806534,
-0.0026142399292439222,
0.021062910556793213,
-0.0331612229347229,
0.01566789112985134,
0... | 0.199454 |
volumes. - list - ``[]`` \* - :spelling:ignore:`forceDeviceDetection` - Forces the auto-detection of devices, even if specific devices are explicitly listed - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.enableAlpn` - Enable ALPN for all listeners configured with Gateway API. ALPN will attempt HTTP/2, then HTTP 1.1. Note that this will also enable ``appProtocol`` support, and services that wish to use HTTP/2 will need to indicate that via their ``appProtocol``. - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.enableAppProtocol` - Enable Backend Protocol selection support (GEP-1911) for Gateway API via appProtocol. - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.enableProxyProtocol` - Enable proxy protocol for all GatewayAPI listeners. Note that \*only\* Proxy protocol traffic will be accepted once this is enabled. - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.enabled` - Enable support for Gateway API in cilium This will automatically set enable-envoy-config as well. - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.externalTrafficPolicy` - Control how traffic from external sources is routed to the LoadBalancer Kubernetes Service for all Cilium GatewayAPI Gateway instances. Valid values are "Cluster" and "Local". Note that this value will be ignored when ``hostNetwork.enabled == true``. ref: https://kubernetes.io/docs/reference/networking/virtual-ips/#external-traffic-policy - string - ``"Cluster"`` \* - :spelling:ignore:`gatewayAPI.gatewayClass.create` - Enable creation of GatewayClass resource The default value is 'auto' which decides according to presence of gateway.networking.k8s.io/v1/GatewayClass in the cluster. Other possible values are 'true' and 'false', which will either always or never create the GatewayClass, respectively. - string - ``"auto"`` \* - :spelling:ignore:`gatewayAPI.hostNetwork.enabled` - Configure whether the Envoy listeners should be exposed on the host network. - bool - ``false`` \* - :spelling:ignore:`gatewayAPI.hostNetwork.nodes.matchLabels` - Specify the labels of the nodes where the Ingress listeners should be exposed matchLabels: kubernetes.io/os: linux kubernetes.io/hostname: kind-worker - object - ``{}`` \* - :spelling:ignore:`gatewayAPI.secretsNamespace` - SecretsNamespace is the namespace in which envoy SDS will retrieve TLS secrets from. - object - ``{"create":true,"name":"cilium-secrets","sync":true}`` \* - :spelling:ignore:`gatewayAPI.secretsNamespace.create` - Create secrets namespace for Gateway API. - bool - ``true`` \* - :spelling:ignore:`gatewayAPI.secretsNamespace.name` - Name of Gateway API secret namespace. - string - ``"cilium-secrets"`` \* - :spelling:ignore:`gatewayAPI.secretsNamespace.sync` - Enable secret sync, which will make sure all TLS secrets used by Ingress are synced to secretsNamespace.name. If disabled, TLS secrets must be maintained externally. - bool - ``true`` \* - :spelling:ignore:`gatewayAPI.xffNumTrustedHops` - The number of additional GatewayAPI proxy hops from the right side of the HTTP header to trust when determining the origin client's IP address. - int - ``0`` \* - :spelling:ignore:`gke.enabled` - Enable Google Kubernetes Engine integration - bool - ``false`` \* - :spelling:ignore:`healthCheckICMPFailureThreshold` - Number of ICMP requests sent for each health check before marking a node or endpoint unreachable. - int - ``3`` \* - :spelling:ignore:`healthChecking` - Enable connectivity health checking. - bool - ``true`` \* - :spelling:ignore:`healthPort` - TCP port for the agent health API. This is not the port for cilium-health. - int - ``9879`` \* - :spelling:ignore:`hostFirewall` - Configure the host firewall. - object - ``{"enabled":false}`` \* - :spelling:ignore:`hostFirewall.enabled` - Enables the enforcement of host policies in the eBPF datapath. - bool - ``false`` \* - :spelling:ignore:`hubble.annotations` - Annotations to be added to all top-level hubble objects (resources under templates/hubble) - object - ``{}`` \* - :spelling:ignore:`hubble.dropEventEmitter` - Emit v1.Events related to pods on detection of packet drops. This feature is alpha, please provide feedback at https://github.com/cilium/cilium/issues/33975. - object - ``{"enabled":false,"interval":"2m","reasons":["auth\_required","policy\_denied"]}`` \* - :spelling:ignore:`hubble.dropEventEmitter.interval` - - Minimum time between emitting same events. - string - ``"2m"`` \* - :spelling:ignore:`hubble.dropEventEmitter.reasons` - - Drop reasons to emit events for. ref: https://docs.cilium.io/en/stable/\_api/v1/flow/README/#dropreason - list - ``["auth\_required","policy\_denied"]`` \* - :spelling:ignore:`hubble.enabled` - Enable Hubble (true by default). - bool - ``true`` \* - :spelling:ignore:`hubble.export` - Hubble flows export. - object - ``{"dynamic":{"config":{"configMapName":"cilium-flowlog-config","content":[{"aggregationInterval":"0s","excludeFilters":[],"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log","includeFilters":[],"name":"all"}],"createConfigMap":true},"enabled":false},"static":{"aggregationInterval":"0s","allowList":[],"denyList":[],"enabled":false,"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log"}}`` \* - :spelling:ignore:`hubble.export.dynamic` - | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.051517993211746216,
-0.003546519670635462,
-0.0011018466902896762,
-0.043321989476680756,
-0.03696313500404358,
-0.07084807008504868,
-0.050201475620269775,
-0.03515397384762764,
-0.011602601036429405,
-0.030152583494782448,
0.046853791922330856,
-0.02508237212896347,
0.01764538697898388,... | 0.001717 |
events. - string - ``"2m"`` \* - :spelling:ignore:`hubble.dropEventEmitter.reasons` - - Drop reasons to emit events for. ref: https://docs.cilium.io/en/stable/\_api/v1/flow/README/#dropreason - list - ``["auth\_required","policy\_denied"]`` \* - :spelling:ignore:`hubble.enabled` - Enable Hubble (true by default). - bool - ``true`` \* - :spelling:ignore:`hubble.export` - Hubble flows export. - object - ``{"dynamic":{"config":{"configMapName":"cilium-flowlog-config","content":[{"aggregationInterval":"0s","excludeFilters":[],"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log","includeFilters":[],"name":"all"}],"createConfigMap":true},"enabled":false},"static":{"aggregationInterval":"0s","allowList":[],"denyList":[],"enabled":false,"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log"}}`` \* - :spelling:ignore:`hubble.export.dynamic` - - Dynamic exporters configuration. Dynamic exporters may be reconfigured without a need of agent restarts. - object - ``{"config":{"configMapName":"cilium-flowlog-config","content":[{"aggregationInterval":"0s","excludeFilters":[],"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log","includeFilters":[],"name":"all"}],"createConfigMap":true},"enabled":false}`` \* - :spelling:ignore:`hubble.export.dynamic.config.configMapName` - -- Name of configmap with configuration that may be altered to reconfigure exporters within a running agents. - string - ``"cilium-flowlog-config"`` \* - :spelling:ignore:`hubble.export.dynamic.config.content` - -- Exporters configuration in YAML format. - list - ``[{"aggregationInterval":"0s","excludeFilters":[],"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log","includeFilters":[],"name":"all"}]`` \* - :spelling:ignore:`hubble.export.dynamic.config.createConfigMap` - -- True if helm installer should create config map. Switch to false if you want to self maintain the file content. - bool - ``true`` \* - :spelling:ignore:`hubble.export.static` - - Static exporter configuration. Static exporter is bound to agent lifecycle. - object - ``{"aggregationInterval":"0s","allowList":[],"denyList":[],"enabled":false,"fieldAggregate":[],"fieldMask":[],"fileCompress":false,"fileMaxBackups":5,"fileMaxSizeMb":10,"filePath":"/var/run/cilium/hubble/events.log"}`` \* - :spelling:ignore:`hubble.export.static.aggregationInterval` - - Defines the interval at which to aggregate before exporting Hubble flows. Aggregation feature is only enabled when fieldAggregate is specified and aggregationInterval > 0s. - string - ``"0s"`` \* - :spelling:ignore:`hubble.export.static.fileCompress` - - Enable compression of rotated files. - bool - ``false`` \* - :spelling:ignore:`hubble.export.static.fileMaxBackups` - - Defines max number of backup/rotated files. - int - ``5`` \* - :spelling:ignore:`hubble.export.static.fileMaxSizeMb` - - Defines max file size of output file before it gets rotated. - int - ``10`` \* - :spelling:ignore:`hubble.listenAddress` - An additional address for Hubble to listen to. Set this field ":4244" if you are enabling Hubble Relay, as it assumes that Hubble is listening on port 4244. - string - ``":4244"`` \* - :spelling:ignore:`hubble.metrics` - Hubble metrics configuration. See https://docs.cilium.io/en/stable/observability/metrics/#hubble-metrics for more comprehensive documentation about Hubble metrics. - object - ``{"dashboards":{"annotations":{},"enabled":false,"label":"grafana\_dashboard","labelValue":"1","namespace":null},"dynamic":{"config":{"configMapName":"cilium-dynamic-metrics-config","content":[],"createConfigMap":true},"enabled":false},"enableOpenMetrics":false,"enabled":null,"port":9965,"serviceAnnotations":{},"serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","jobLabel":"","labels":{},"metricRelabelings":null,"relabelings":[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}],"scrapeTimeout":null,"tlsConfig":{}},"tls":{"enabled":false,"server":{"cert":"","existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":"","mtls":{"enabled":false,"key":"ca.crt","name":null,"useSecret":false}}}}`` \* - :spelling:ignore:`hubble.metrics.dashboards` - Grafana dashboards for hubble grafana can import dashboards based on the label and value ref: https://github.com/grafana/helm-charts/tree/main/charts/grafana#sidecar-for-dashboards - object - ``{"annotations":{},"enabled":false,"label":"grafana\_dashboard","labelValue":"1","namespace":null}`` \* - :spelling:ignore:`hubble.metrics.dynamic.config.configMapName` - -- Name of configmap with configuration that may be altered to reconfigure metric handlers within a running agent. - string - ``"cilium-dynamic-metrics-config"`` \* - :spelling:ignore:`hubble.metrics.dynamic.config.content` - -- Exporters configuration in YAML format. - list - ``[]`` \* - :spelling:ignore:`hubble.metrics.dynamic.config.createConfigMap` - -- True if helm installer should create config map. Switch to false if you want to self maintain the file content. - bool - ``true`` \* - :spelling:ignore:`hubble.metrics.enableOpenMetrics` - Enables exporting hubble metrics in OpenMetrics format. - bool - ``false`` \* - :spelling:ignore:`hubble.metrics.enabled` - Configures the list of metrics to collect. If empty or null, metrics are disabled. Example: enabled: - dns:query;ignoreAAAA - drop - tcp - flow - icmp - http You can specify the list of metrics from the helm CLI: --set hubble.metrics.enabled="{dns:query;ignoreAAAA,drop,tcp,flow,icmp,http}" - string - ``nil`` \* - :spelling:ignore:`hubble.metrics.port` - Configure the port the hubble metric server listens on. - int - ``9965`` \* - :spelling:ignore:`hubble.metrics.serviceAnnotations` - Annotations to be added to hubble-metrics service. - object - ``{}`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.annotations` - Annotations to add to ServiceMonitor hubble - object - ``{}`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.enabled` - Create ServiceMonitor resources for Prometheus Operator. This requires the prometheus CRDs to be available. ref: https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.interval` - Interval for scrape metrics. - string - ``"10s"`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.jobLabel` - jobLabel to add for ServiceMonitor hubble - string - ``""`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.labels` - Labels to add to ServiceMonitor hubble - object - ``{}`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor hubble - string - ``nil`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor hubble - list - ``[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}]`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.026512697339057922,
0.013020885176956654,
-0.03750154748558998,
0.03206547349691391,
0.0770975723862648,
-0.02468469925224781,
-0.03810394927859306,
0.046205442398786545,
0.06377585977315903,
-0.030677543953061104,
-0.02146117202937603,
-0.11474926769733429,
-0.0396193228662014,
-0.00377... | 0.131546 |
- Labels to add to ServiceMonitor hubble - object - ``{}`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor hubble - string - ``nil`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor hubble - list - ``[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}]`` \* - :spelling:ignore:`hubble.metrics.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`hubble.metrics.tls.server.cert` - base64 encoded PEM values for the Hubble metrics server certificate (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.metrics.tls.server.existingSecret` - Name of the Secret containing the certificate and key for the Hubble metrics server. If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.metrics.tls.server.extraDnsNames` - Extra DNS names added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`hubble.metrics.tls.server.extraIpAddresses` - Extra IP addresses added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`hubble.metrics.tls.server.key` - base64 encoded PEM values for the Hubble metrics server key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.metrics.tls.server.mtls` - Configure mTLS for the Hubble metrics server. - object - ``{"enabled":false,"key":"ca.crt","name":null,"useSecret":false}`` \* - :spelling:ignore:`hubble.metrics.tls.server.mtls.key` - Entry of the ConfigMap containing the CA. - string - ``"ca.crt"`` \* - :spelling:ignore:`hubble.metrics.tls.server.mtls.name` - Name of the ConfigMap containing the CA to validate client certificates against. If mTLS is enabled and this is unspecified, it will default to the same CA used for Hubble metrics server certificates. - string - ``nil`` \* - :spelling:ignore:`hubble.networkPolicyCorrelation` - Enables network policy correlation of Hubble flows, i.e. populating ``egress\_allowed\_by``\ , ``ingress\_denied\_by`` fields with policy information. - object - ``{"enabled":true}`` \* - :spelling:ignore:`hubble.peerService.clusterDomain` - The cluster domain to use to query the Hubble Peer service. It should be the local cluster. - string - ``"cluster.local"`` \* - :spelling:ignore:`hubble.peerService.targetPort` - Target Port for the Peer service, must match the hubble.listenAddress' port. - int - ``4244`` \* - :spelling:ignore:`hubble.preferIpv6` - Whether Hubble should prefer to announce IPv6 or IPv4 addresses if both are available. - bool - ``false`` \* - :spelling:ignore:`hubble.redact` - Enables redacting sensitive information present in Layer 7 flows. - object - ``{"enabled":false,"http":{"headers":{"allow":[],"deny":[]},"urlQuery":false,"userInfo":true},"kafka":{"apiKey":true}}`` \* - :spelling:ignore:`hubble.redact.http.headers.allow` - List of HTTP headers to allow: headers not matching will be redacted. Note: ``allow`` and ``deny`` lists cannot be used both at the same time, only one can be present. Example: redact: enabled: true http: headers: allow: - traceparent - tracestate - Cache-Control You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.http.headers.allow="traceparent,tracestate,Cache-Control" - list - ``[]`` \* - :spelling:ignore:`hubble.redact.http.headers.deny` - List of HTTP headers to deny: matching headers will be redacted. Note: ``allow`` and ``deny`` lists cannot be used both at the same time, only one can be present. Example: redact: enabled: true http: headers: deny: - Authorization - Proxy-Authorization You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.http.headers.deny="Authorization,Proxy-Authorization" - list - ``[]`` \* - :spelling:ignore:`hubble.redact.http.urlQuery` - Enables redacting URL query (GET) parameters. Example: redact: enabled: true http: urlQuery: true You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.http.urlQuery="true" - bool - ``false`` \* - :spelling:ignore:`hubble.redact.http.userInfo` - Enables redacting user info, e.g., password when basic auth is used. Example: redact: enabled: true http: userInfo: true You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.http.userInfo="true" - bool - ``true`` \* - :spelling:ignore:`hubble.redact.kafka.apiKey` - Enables redacting Kafka's API key (deprecated, will be removed in v1.19). Example: redact: enabled: true kafka: apiKey: true You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.kafka.apiKey="true" - bool - ``true`` \* - :spelling:ignore:`hubble.relay.affinity` - Affinity for hubble-replay - object - ``{"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`hubble.relay.annotations` - Annotations to be added | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.01682269759476185,
0.021798761561512947,
-0.0293886736035347,
0.02689352072775364,
0.062360577285289764,
0.02959173358976841,
0.029541542753577232,
-0.0398661345243454,
0.01412851084023714,
0.002314566867426038,
0.022638054564595222,
-0.1046856939792633,
0.032563116401433945,
-0.05243100... | 0.078074 |
key (deprecated, will be removed in v1.19). Example: redact: enabled: true kafka: apiKey: true You can specify the options from the helm CLI: --set hubble.redact.enabled="true" --set hubble.redact.kafka.apiKey="true" - bool - ``true`` \* - :spelling:ignore:`hubble.relay.affinity` - Affinity for hubble-replay - object - ``{"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`hubble.relay.annotations` - Annotations to be added to all top-level hubble-relay objects (resources under templates/hubble-relay) - object - ``{}`` \* - :spelling:ignore:`hubble.relay.enabled` - Enable Hubble Relay (requires hubble.enabled=true) - bool - ``false`` \* - :spelling:ignore:`hubble.relay.extraEnv` - Additional hubble-relay environment variables. - list - ``[]`` \* - :spelling:ignore:`hubble.relay.extraVolumeMounts` - Additional hubble-relay volumeMounts. - list - ``[]`` \* - :spelling:ignore:`hubble.relay.extraVolumes` - Additional hubble-relay volumes. - list - ``[]`` \* - :spelling:ignore:`hubble.relay.gops.enabled` - Enable gops for hubble-relay - bool - ``true`` \* - :spelling:ignore:`hubble.relay.gops.port` - Configure gops listen port for hubble-relay - int - ``9893`` \* - :spelling:ignore:`hubble.relay.image` - Hubble-relay container image. - object - ``{"digest":"","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/hubble-relay-ci","tag":"latest","useDigest":false}`` \* - :spelling:ignore:`hubble.relay.listenHost` - Host to listen to. Specify an empty string to bind to all the interfaces. - string - ``""`` \* - :spelling:ignore:`hubble.relay.listenPort` - Port to listen to. - string - ``"4245"`` \* - :spelling:ignore:`hubble.relay.nodeSelector` - Node labels for pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`hubble.relay.podAnnotations` - Annotations to be added to hubble-relay pods - object - ``{}`` \* - :spelling:ignore:`hubble.relay.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`hubble.relay.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`hubble.relay.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by ``maxUnavailable: null`` - string - ``nil`` \* - :spelling:ignore:`hubble.relay.podDisruptionBudget.unhealthyPodEvictionPolicy` - How are unhealthy, but running, pods counted for eviction - string - ``nil`` \* - :spelling:ignore:`hubble.relay.podLabels` - Labels to be added to hubble-relay pods - object - ``{}`` \* - :spelling:ignore:`hubble.relay.podSecurityContext` - hubble-relay pod security context - object - ``{"fsGroup":65532,"seccompProfile":{"type":"RuntimeDefault"}}`` \* - :spelling:ignore:`hubble.relay.pprof.address` - Configure pprof listen address for hubble-relay - string - ``"localhost"`` \* - :spelling:ignore:`hubble.relay.pprof.blockProfileRate` - Enable goroutine blocking profiling for hubble-relay and set the rate of sampled events in nanoseconds (set to 1 to sample all events [warning: performance overhead]) - int - ``0`` \* - :spelling:ignore:`hubble.relay.pprof.enabled` - Enable pprof for hubble-relay - bool - ``false`` \* - :spelling:ignore:`hubble.relay.pprof.mutexProfileFraction` - Enable mutex contention profiling for hubble-relay and set the fraction of sampled events (set to 1 to sample all events) - int - ``0`` \* - :spelling:ignore:`hubble.relay.pprof.port` - Configure pprof listen port for hubble-relay - int - ``6062`` \* - :spelling:ignore:`hubble.relay.priorityClassName` - The priority class to use for hubble-relay - string - ``""`` \* - :spelling:ignore:`hubble.relay.prometheus` - Enable prometheus metrics for hubble-relay on the configured port at /metrics - object - ``{"enabled":false,"port":9966,"serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","labels":{},"metricRelabelings":null,"relabelings":null,"scrapeTimeout":null}}`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.annotations` - Annotations to add to ServiceMonitor hubble-relay - object - ``{}`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.enabled` - Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.interval` - Interval for scrape metrics. - string - ``"10s"`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.labels` - Labels to add to ServiceMonitor hubble-relay - object - ``{}`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor hubble-relay - string - ``nil`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor hubble-relay - string - ``nil`` \* - :spelling:ignore:`hubble.relay.prometheus.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`hubble.relay.replicas` - Number of replicas run for the hubble-relay deployment. - int - ``1`` \* - :spelling:ignore:`hubble.relay.resources` - Specifies the resources for the hubble-relay pods - object - ``{}`` \* - :spelling:ignore:`hubble.relay.retryTimeout` - Backoff duration to retry connecting to the local hubble instance in case of failure | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.01126294769346714,
0.03429042175412178,
-0.017679428681731224,
0.033142123371362686,
0.05035039782524109,
0.08407767862081528,
-0.027578137814998627,
-0.015851309522986412,
0.03816450387239456,
0.05844389647245407,
-0.016277341172099113,
-0.03682778403162956,
0.018096717074513435,
-0.035... | 0.10336 |
``nil`` \* - :spelling:ignore:`hubble.relay.replicas` - Number of replicas run for the hubble-relay deployment. - int - ``1`` \* - :spelling:ignore:`hubble.relay.resources` - Specifies the resources for the hubble-relay pods - object - ``{}`` \* - :spelling:ignore:`hubble.relay.retryTimeout` - Backoff duration to retry connecting to the local hubble instance in case of failure (e.g. "30s"). - string - ``nil`` \* - :spelling:ignore:`hubble.relay.rollOutPods` - Roll out Hubble Relay pods automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`hubble.relay.securityContext` - hubble-relay container security context - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]},"runAsGroup":65532,"runAsNonRoot":true,"runAsUser":65532,"seccompProfile":{"type":"RuntimeDefault"}}`` \* - :spelling:ignore:`hubble.relay.service` - hubble-relay service configuration. - object - ``{"nodePort":31234,"type":"ClusterIP"}`` \* - :spelling:ignore:`hubble.relay.service.nodePort` - - The port to use when the service type is set to NodePort. - int - ``31234`` \* - :spelling:ignore:`hubble.relay.service.type` - - The type of service used for Hubble Relay access, either ClusterIP, NodePort or LoadBalancer. - string - ``"ClusterIP"`` \* - :spelling:ignore:`hubble.relay.sortBufferDrainTimeout` - When the per-request flows sort buffer is not full, a flow is drained every time this timeout is reached (only affects requests in follow-mode) (e.g. "1s"). - string - ``nil`` \* - :spelling:ignore:`hubble.relay.sortBufferLenMax` - Max number of flows that can be buffered for sorting before being sent to the client (per request) (e.g. 100). - int - ``nil`` \* - :spelling:ignore:`hubble.relay.terminationGracePeriodSeconds` - Configure termination grace period for hubble relay Deployment. - int - ``1`` \* - :spelling:ignore:`hubble.relay.tls` - TLS configuration for Hubble Relay - object - ``{"client":{"cert":"","existingSecret":"","key":""},"server":{"cert":"","enabled":false,"existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":"","mtls":false,"relayName":"ui.hubble-relay.cilium.io"}}`` \* - :spelling:ignore:`hubble.relay.tls.client` - The hubble-relay client certificate and private key. This keypair is presented to Hubble server instances for mTLS authentication and is required when hubble.tls.enabled is true. These values need to be set manually if hubble.tls.auto.enabled is false. - object - ``{"cert":"","existingSecret":"","key":""}`` \* - :spelling:ignore:`hubble.relay.tls.client.cert` - base64 encoded PEM values for the Hubble relay client certificate (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tls.client.existingSecret` - Name of the Secret containing the certificate and key for the Hubble metrics server. If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tls.client.key` - base64 encoded PEM values for the Hubble relay client key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tls.server` - The hubble-relay server certificate and private key - object - ``{"cert":"","enabled":false,"existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":"","mtls":false,"relayName":"ui.hubble-relay.cilium.io"}`` \* - :spelling:ignore:`hubble.relay.tls.server.cert` - base64 encoded PEM values for the Hubble relay server certificate (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tls.server.existingSecret` - Name of the Secret containing the certificate and key for the Hubble relay server. If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tls.server.extraDnsNames` - extra DNS names added to certificate when its auto gen - list - ``[]`` \* - :spelling:ignore:`hubble.relay.tls.server.extraIpAddresses` - extra IP addresses added to certificate when its auto gen - list - ``[]`` \* - :spelling:ignore:`hubble.relay.tls.server.key` - base64 encoded PEM values for the Hubble relay server key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.relay.tolerations` - Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`hubble.relay.topologySpreadConstraints` - Pod topology spread constraints for hubble-relay - list - ``[]`` \* - :spelling:ignore:`hubble.relay.updateStrategy` - hubble-relay update strategy - object - ``{"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"}`` \* - :spelling:ignore:`hubble.skipUnknownCGroupIDs` - Skip Hubble events with unknown cgroup ids - bool - ``true`` \* - :spelling:ignore:`hubble.socketPath` - Unix domain socket path to listen to when Hubble is enabled. - string - ``"/var/run/cilium/hubble.sock"`` \* - :spelling:ignore:`hubble.tls` - TLS configuration for Hubble - object - ``{"auto":{"certManagerIssuerRef":{},"certValidityDuration":365,"enabled":true,"method":"helm","schedule":"0 0 1 \*/4 \*"},"enabled":true,"server":{"cert":"","existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":""}}`` \* - :spelling:ignore:`hubble.tls.auto` - Configure automatic TLS certificates generation. - object - ``{"certManagerIssuerRef":{},"certValidityDuration":365,"enabled":true,"method":"helm","schedule":"0 0 1 \*/4 \*"}`` \* - :spelling:ignore:`hubble.tls.auto.certManagerIssuerRef` - certmanager issuer used when hubble.tls.auto.method=certmanager. - object - ``{}`` \* - :spelling:ignore:`hubble.tls.auto.certValidityDuration` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.034565117210149765,
0.010046780109405518,
0.003392402548342943,
0.05727916583418846,
0.07421869039535522,
-0.012979840859770775,
-0.04040726274251938,
-0.03757946938276291,
-0.02909276634454727,
0.042868588119745255,
0.005483394488692284,
0.016748247668147087,
0.060304950922727585,
-0.01... | 0.143503 |
``"/var/run/cilium/hubble.sock"`` \* - :spelling:ignore:`hubble.tls` - TLS configuration for Hubble - object - ``{"auto":{"certManagerIssuerRef":{},"certValidityDuration":365,"enabled":true,"method":"helm","schedule":"0 0 1 \*/4 \*"},"enabled":true,"server":{"cert":"","existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":""}}`` \* - :spelling:ignore:`hubble.tls.auto` - Configure automatic TLS certificates generation. - object - ``{"certManagerIssuerRef":{},"certValidityDuration":365,"enabled":true,"method":"helm","schedule":"0 0 1 \*/4 \*"}`` \* - :spelling:ignore:`hubble.tls.auto.certManagerIssuerRef` - certmanager issuer used when hubble.tls.auto.method=certmanager. - object - ``{}`` \* - :spelling:ignore:`hubble.tls.auto.certValidityDuration` - Generated certificates validity duration in days. Defaults to 365 days (1 year) because MacOS does not accept self-signed certificates with expirations > 825 days. - int - ``365`` \* - :spelling:ignore:`hubble.tls.auto.enabled` - Auto-generate certificates. When set to true, automatically generate a CA and certificates to enable mTLS between Hubble server and Hubble Relay instances. If set to false, the certs for Hubble server need to be provided by setting appropriate values below. - bool - ``true`` \* - :spelling:ignore:`hubble.tls.auto.method` - Set the method to auto-generate certificates. Supported values: - helm: This method uses Helm to generate all certificates. - cronJob: This method uses a Kubernetes CronJob the generate any certificates not provided by the user at installation time. - certmanager: This method use cert-manager to generate & rotate certificates. - string - ``"helm"`` \* - :spelling:ignore:`hubble.tls.auto.schedule` - Schedule for certificates regeneration (regardless of their expiration date). Only used if method is "cronJob". If nil, then no recurring job will be created. Instead, only the one-shot job is deployed to generate the certificates at installation time. Defaults to midnight of the first day of every fourth month. For syntax, see https://kubernetes.io/docs/concepts/workloads/controllers/cron-jobs/#schedule-syntax - string - ``"0 0 1 \*/4 \*"`` \* - :spelling:ignore:`hubble.tls.enabled` - Enable mutual TLS for listenAddress. Setting this value to false is highly discouraged as the Hubble API provides access to potentially sensitive network flow metadata and is exposed on the host network. - bool - ``true`` \* - :spelling:ignore:`hubble.tls.server` - The Hubble server certificate and private key - object - ``{"cert":"","existingSecret":"","extraDnsNames":[],"extraIpAddresses":[],"key":""}`` \* - :spelling:ignore:`hubble.tls.server.cert` - base64 encoded PEM values for the Hubble server certificate (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.tls.server.existingSecret` - Name of the Secret containing the certificate and key for the Hubble server. If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.tls.server.extraDnsNames` - Extra DNS names added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`hubble.tls.server.extraIpAddresses` - Extra IP addresses added to certificate when it's auto generated - list - ``[]`` \* - :spelling:ignore:`hubble.tls.server.key` - base64 encoded PEM values for the Hubble server key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.ui.affinity` - Affinity for hubble-ui - object - ``{}`` \* - :spelling:ignore:`hubble.ui.annotations` - Annotations to be added to all top-level hubble-ui objects (resources under templates/hubble-ui) - object - ``{}`` \* - :spelling:ignore:`hubble.ui.backend.extraEnv` - Additional hubble-ui backend environment variables. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.backend.extraVolumeMounts` - Additional hubble-ui backend volumeMounts. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.backend.extraVolumes` - Additional hubble-ui backend volumes. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.backend.image` - Hubble-ui backend image. - object - ``{"digest":"sha256:db1454e45dc39ca41fbf7cad31eec95d99e5b9949c39daaad0fa81ef29d56953","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/hubble-ui-backend","tag":"v0.13.3","useDigest":true}`` \* - :spelling:ignore:`hubble.ui.backend.resources` - Resource requests and limits for the 'backend' container of the 'hubble-ui' deployment. - object - ``{}`` \* - :spelling:ignore:`hubble.ui.backend.securityContext` - Hubble-ui backend security context. - object - ``{"allowPrivilegeEscalation":false}`` \* - :spelling:ignore:`hubble.ui.baseUrl` - Defines base url prefix for all hubble-ui http requests. It needs to be changed in case if ingress for hubble-ui is configured under some sub-path. Trailing ``/`` is required for custom path, ex. ``/service-map/`` - string - ``"/"`` \* - :spelling:ignore:`hubble.ui.enabled` - Whether to enable the Hubble UI. - bool - ``false`` \* - :spelling:ignore:`hubble.ui.frontend.extraEnv` - Additional hubble-ui frontend environment variables. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.frontend.extraVolumeMounts` - Additional hubble-ui frontend volumeMounts. - list | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.0049544693902134895,
0.07625913619995117,
-0.05155058950185776,
-0.01729256473481655,
0.017429428175091743,
-0.03242695331573486,
-0.06013800948858261,
0.011348510161042213,
0.017362752929329872,
-0.004676789976656437,
0.01525139156728983,
-0.053379807621240616,
0.0625552162528038,
0.02... | 0.002349 |
Trailing ``/`` is required for custom path, ex. ``/service-map/`` - string - ``"/"`` \* - :spelling:ignore:`hubble.ui.enabled` - Whether to enable the Hubble UI. - bool - ``false`` \* - :spelling:ignore:`hubble.ui.frontend.extraEnv` - Additional hubble-ui frontend environment variables. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.frontend.extraVolumeMounts` - Additional hubble-ui frontend volumeMounts. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.frontend.extraVolumes` - Additional hubble-ui frontend volumes. - list - ``[]`` \* - :spelling:ignore:`hubble.ui.frontend.image` - Hubble-ui frontend image. - object - ``{"digest":"sha256:661d5de7050182d495c6497ff0b007a7a1e379648e60830dd68c4d78ae21761d","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/hubble-ui","tag":"v0.13.3","useDigest":true}`` \* - :spelling:ignore:`hubble.ui.frontend.resources` - Resource requests and limits for the 'frontend' container of the 'hubble-ui' deployment. - object - ``{}`` \* - :spelling:ignore:`hubble.ui.frontend.securityContext` - Hubble-ui frontend security context. - object - ``{"allowPrivilegeEscalation":false}`` \* - :spelling:ignore:`hubble.ui.frontend.server.ipv6` - Controls server listener for ipv6 - object - ``{"enabled":true}`` \* - :spelling:ignore:`hubble.ui.ingress` - hubble-ui ingress configuration. - object - ``{"annotations":{},"className":"","enabled":false,"hosts":["chart-example.local"],"labels":{},"tls":[]}`` \* - :spelling:ignore:`hubble.ui.labels` - Additional labels to be added to 'hubble-ui' deployment object - object - ``{}`` \* - :spelling:ignore:`hubble.ui.nodeSelector` - Node labels for pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`hubble.ui.podAnnotations` - Annotations to be added to hubble-ui pods - object - ``{}`` \* - :spelling:ignore:`hubble.ui.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`hubble.ui.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`hubble.ui.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by ``maxUnavailable: null`` - string - ``nil`` \* - :spelling:ignore:`hubble.ui.podDisruptionBudget.unhealthyPodEvictionPolicy` - How are unhealthy, but running, pods counted for eviction - string - ``nil`` \* - :spelling:ignore:`hubble.ui.podLabels` - Labels to be added to hubble-ui pods - object - ``{}`` \* - :spelling:ignore:`hubble.ui.priorityClassName` - The priority class to use for hubble-ui - string - ``""`` \* - :spelling:ignore:`hubble.ui.replicas` - The number of replicas of Hubble UI to deploy. - int - ``1`` \* - :spelling:ignore:`hubble.ui.rollOutPods` - Roll out Hubble-ui pods automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`hubble.ui.securityContext` - Security context to be added to Hubble UI pods - object - ``{"fsGroup":1001,"runAsGroup":1001,"runAsUser":1001}`` \* - :spelling:ignore:`hubble.ui.service` - hubble-ui service configuration. - object - ``{"annotations":{},"labels":{},"nodePort":31235,"type":"ClusterIP"}`` \* - :spelling:ignore:`hubble.ui.service.annotations` - Annotations to be added for the Hubble UI service - object - ``{}`` \* - :spelling:ignore:`hubble.ui.service.labels` - Labels to be added for the Hubble UI service - object - ``{}`` \* - :spelling:ignore:`hubble.ui.service.nodePort` - - The port to use when the service type is set to NodePort. - int - ``31235`` \* - :spelling:ignore:`hubble.ui.service.type` - - The type of service used for Hubble UI access, either ClusterIP or NodePort. - string - ``"ClusterIP"`` \* - :spelling:ignore:`hubble.ui.standalone.enabled` - When true, it will allow installing the Hubble UI only, without checking dependencies. It is useful if a cluster already has cilium and Hubble relay installed and you just want Hubble UI to be deployed. When installed via helm, installing UI should be done via ``helm upgrade`` and when installed via the cilium cli, then ``cilium hubble enable --ui`` - bool - ``false`` \* - :spelling:ignore:`hubble.ui.standalone.tls.certsVolume` - When deploying Hubble UI in standalone, with tls enabled for Hubble relay, it is required to provide a volume for mounting the client certificates. - object - ``{}`` \* - :spelling:ignore:`hubble.ui.tls.client.cert` - base64 encoded PEM values for the Hubble UI client certificate (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.ui.tls.client.existingSecret` - Name of the Secret containing the client certificate and key for Hubble UI If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.ui.tls.client.key` - base64 encoded PEM values for the Hubble UI client key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.ui.tmpVolume` - Configure temporary volume for hubble-ui - | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.09770562499761581,
-0.01637767255306244,
-0.005859097931534052,
-0.00047826595255173743,
0.07777676731348038,
0.018598150461912155,
-0.06480136513710022,
0.10203427821397781,
-0.02798757143318653,
-0.02301369607448578,
0.05524006858468056,
-0.054086100310087204,
0.0523240901529789,
0.032... | -0.030388 |
client certificate and key for Hubble UI If specified, cert and key are ignored. - string - ``""`` \* - :spelling:ignore:`hubble.ui.tls.client.key` - base64 encoded PEM values for the Hubble UI client key (deprecated). Use existingSecret instead. - string - ``""`` \* - :spelling:ignore:`hubble.ui.tmpVolume` - Configure temporary volume for hubble-ui - object - ``{}`` \* - :spelling:ignore:`hubble.ui.tolerations` - Node tolerations for pod assignment on nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[]`` \* - :spelling:ignore:`hubble.ui.topologySpreadConstraints` - Pod topology spread constraints for hubble-ui - list - ``[]`` \* - :spelling:ignore:`hubble.ui.updateStrategy` - hubble-ui update strategy. - object - ``{"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"}`` \* - :spelling:ignore:`identityAllocationMode` - Method to use for identity allocation (\ ``crd``\ , ``kvstore`` or ``doublewrite-readkvstore`` / ``doublewrite-readcrd`` for migrating between identity backends). - string - ``"crd"`` \* - :spelling:ignore:`identityChangeGracePeriod` - Time to wait before using new identity on endpoint identity change. - string - ``"5s"`` \* - :spelling:ignore:`identityManagementMode` - Control whether CiliumIdentities are created by the agent ("agent"), the operator ("operator") or both ("both"). "Both" should be used only to migrate between "agent" and "operator". Operator-managed identities is a beta feature. - string - ``"agent"`` \* - :spelling:ignore:`image` - Agent container image. - object - ``{"digest":"","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/cilium-ci","tag":"latest","useDigest":false}`` \* - :spelling:ignore:`imagePullSecrets` - Configure image pull secrets for pulling container images - list - ``[]`` \* - :spelling:ignore:`ingressController.default` - Set cilium ingress controller to be the default ingress controller This will let cilium ingress controller route entries without ingress class set - bool - ``false`` \* - :spelling:ignore:`ingressController.defaultSecretName` - Default secret name for ingresses without .spec.tls[].secretName set. - string - ``nil`` \* - :spelling:ignore:`ingressController.defaultSecretNamespace` - Default secret namespace for ingresses without .spec.tls[].secretName set. - string - ``nil`` \* - :spelling:ignore:`ingressController.enableProxyProtocol` - Enable proxy protocol for all Ingress listeners. Note that \*only\* Proxy protocol traffic will be accepted once this is enabled. - bool - ``false`` \* - :spelling:ignore:`ingressController.enabled` - Enable cilium ingress controller This will automatically set enable-envoy-config as well. - bool - ``false`` \* - :spelling:ignore:`ingressController.enforceHttps` - Enforce https for host having matching TLS host in Ingress. Incoming traffic to http listener will return 308 http error code with respective location in header. - bool - ``true`` \* - :spelling:ignore:`ingressController.hostNetwork.enabled` - Configure whether the Envoy listeners should be exposed on the host network. - bool - ``false`` \* - :spelling:ignore:`ingressController.hostNetwork.nodes.matchLabels` - Specify the labels of the nodes where the Ingress listeners should be exposed matchLabels: kubernetes.io/os: linux kubernetes.io/hostname: kind-worker - object - ``{}`` \* - :spelling:ignore:`ingressController.hostNetwork.sharedListenerPort` - Configure a specific port on the host network that gets used for the shared listener. - int - ``8080`` \* - :spelling:ignore:`ingressController.ingressLBAnnotationPrefixes` - IngressLBAnnotations are the annotation and label prefixes, which are used to filter annotations and/or labels to propagate from Ingress to the Load Balancer service - list - ``["lbipam.cilium.io","nodeipam.cilium.io","service.beta.kubernetes.io","service.kubernetes.io","cloud.google.com"]`` \* - :spelling:ignore:`ingressController.loadbalancerMode` - Default ingress load balancer mode Supported values: shared, dedicated For granular control, use the following annotations on the ingress resource: "ingress.cilium.io/loadbalancer-mode: dedicated" (or "shared"). - string - ``"dedicated"`` \* - :spelling:ignore:`ingressController.secretsNamespace` - SecretsNamespace is the namespace in which envoy SDS will retrieve TLS secrets from. - object - ``{"create":true,"name":"cilium-secrets","sync":true}`` \* - :spelling:ignore:`ingressController.secretsNamespace.create` - Create secrets namespace for Ingress. - bool - ``true`` \* - :spelling:ignore:`ingressController.secretsNamespace.name` - Name of Ingress secret namespace. - string - ``"cilium-secrets"`` \* - :spelling:ignore:`ingressController.secretsNamespace.sync` - Enable secret sync, which will make sure all TLS secrets used by Ingress are synced to secretsNamespace.name. If disabled, TLS secrets must be maintained externally. - bool - ``true`` \* - :spelling:ignore:`ingressController.service` - Load-balancer service in shared mode. This is a single load-balancer service for all Ingress resources. - object - ``{"allocateLoadBalancerNodePorts":null,"annotations":{},"externalTrafficPolicy":"Cluster","insecureNodePort":null,"labels":{},"loadBalancerClass":null,"loadBalancerIP":null,"name":"cilium-ingress","secureNodePort":null,"type":"LoadBalancer"}`` \* - :spelling:ignore:`ingressController.service.allocateLoadBalancerNodePorts` - Configure if node port allocation is required | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.013162950053811073,
0.055062707513570786,
0.026422563940286636,
-0.004858311731368303,
0.038793884217739105,
0.0004578556108754128,
-0.014790527522563934,
0.0151148596778512,
0.005893188528716564,
-0.010450044646859169,
0.01568947732448578,
-0.0833008661866188,
0.11923468112945557,
0.019... | -0.020008 |
are synced to secretsNamespace.name. If disabled, TLS secrets must be maintained externally. - bool - ``true`` \* - :spelling:ignore:`ingressController.service` - Load-balancer service in shared mode. This is a single load-balancer service for all Ingress resources. - object - ``{"allocateLoadBalancerNodePorts":null,"annotations":{},"externalTrafficPolicy":"Cluster","insecureNodePort":null,"labels":{},"loadBalancerClass":null,"loadBalancerIP":null,"name":"cilium-ingress","secureNodePort":null,"type":"LoadBalancer"}`` \* - :spelling:ignore:`ingressController.service.allocateLoadBalancerNodePorts` - Configure if node port allocation is required for LB service ref: https://kubernetes.io/docs/concepts/services-networking/service/#load-balancer-nodeport-allocation - string - ``nil`` \* - :spelling:ignore:`ingressController.service.annotations` - Annotations to be added for the shared LB service - object - ``{}`` \* - :spelling:ignore:`ingressController.service.externalTrafficPolicy` - Control how traffic from external sources is routed to the LoadBalancer Kubernetes Service for Cilium Ingress in shared mode. Valid values are "Cluster" and "Local". ref: https://kubernetes.io/docs/reference/networking/virtual-ips/#external-traffic-policy - string - ``"Cluster"`` \* - :spelling:ignore:`ingressController.service.insecureNodePort` - Configure a specific nodePort for insecure HTTP traffic on the shared LB service - string - ``nil`` \* - :spelling:ignore:`ingressController.service.labels` - Labels to be added for the shared LB service - object - ``{}`` \* - :spelling:ignore:`ingressController.service.loadBalancerClass` - Configure a specific loadBalancerClass on the shared LB service (requires Kubernetes 1.24+) - string - ``nil`` \* - :spelling:ignore:`ingressController.service.loadBalancerIP` - Configure a specific loadBalancerIP on the shared LB service - string - ``nil`` \* - :spelling:ignore:`ingressController.service.name` - Service name - string - ``"cilium-ingress"`` \* - :spelling:ignore:`ingressController.service.secureNodePort` - Configure a specific nodePort for secure HTTPS traffic on the shared LB service - string - ``nil`` \* - :spelling:ignore:`ingressController.service.type` - Service type for the shared LB service - string - ``"LoadBalancer"`` \* - :spelling:ignore:`initResources` - resources & limits for the agent init containers - object - ``{}`` \* - :spelling:ignore:`installNoConntrackIptablesRules` - Install Iptables rules to skip netfilter connection tracking on all pod traffic. This option is only effective when Cilium is running in direct routing and full KPR mode. Moreover, this option cannot be enabled when Cilium is running in a managed Kubernetes environment or in a chained CNI setup. - bool - ``false`` \* - :spelling:ignore:`ipMasqAgent` - Configure the eBPF-based ip-masq-agent - object - ``{"enabled":false}`` \* - :spelling:ignore:`ipam.ciliumNodeUpdateRate` - Maximum rate at which the CiliumNode custom resource is updated. - string - ``"15s"`` \* - :spelling:ignore:`ipam.installUplinkRoutesForDelegatedIPAM` - Install ingress/egress routes through uplink on host for Pods when working with delegated IPAM plugin. - bool - ``false`` \* - :spelling:ignore:`ipam.mode` - Configure IP Address Management mode. ref: https://docs.cilium.io/en/stable/network/concepts/ipam/ - string - ``"cluster-pool"`` \* - :spelling:ignore:`ipam.multiPoolPreAllocation` - Pre-allocation settings for IPAM in Multi-Pool mode - string - ``""`` \* - :spelling:ignore:`ipam.nodeSpec` - NodeSpec configuration for the IPAM - object - ``{"ipamMaxAllocate":null,"ipamMinAllocate":null,"ipamPreAllocate":null,"ipamStaticIPTags":[]}`` \* - :spelling:ignore:`ipam.nodeSpec.ipamMaxAllocate` - IPAM max allocate @schema type: [null, integer] @schema - string - ``nil`` \* - :spelling:ignore:`ipam.nodeSpec.ipamMinAllocate` - IPAM min allocate @schema type: [null, integer] @schema - string - ``nil`` \* - :spelling:ignore:`ipam.nodeSpec.ipamPreAllocate` - IPAM pre allocate @schema type: [null, integer] @schema - string - ``nil`` \* - :spelling:ignore:`ipam.nodeSpec.ipamStaticIPTags` - IPAM static IP tags (currently only works with AWS and Azure) - list - ``[]`` \* - :spelling:ignore:`ipam.operator.autoCreateCiliumPodIPPools` - IP pools to auto-create in multi-pool IPAM mode. - object - ``{}`` \* - :spelling:ignore:`ipam.operator.clusterPoolIPv4MaskSize` - IPv4 CIDR mask size to delegate to individual nodes for IPAM. - int - ``24`` \* - :spelling:ignore:`ipam.operator.clusterPoolIPv4PodCIDRList` - IPv4 CIDR list range to delegate to individual nodes for IPAM. - list - ``["10.0.0.0/8"]`` \* - :spelling:ignore:`ipam.operator.clusterPoolIPv6MaskSize` - IPv6 CIDR mask size to delegate to individual nodes for IPAM. - int - ``120`` \* - :spelling:ignore:`ipam.operator.clusterPoolIPv6PodCIDRList` - IPv6 CIDR list range to delegate to individual nodes for IPAM. - list - ``["fd00::/104"]`` \* - :spelling:ignore:`ipam.operator.externalAPILimitBurstSize` - The maximum burst size when rate limiting access to external APIs. Also known as the token bucket capacity. - int - ``20`` \* - :spelling:ignore:`ipam.operator.externalAPILimitQPS` - The maximum queries per second when rate limiting access to | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.06971476972103119,
0.01859525963664055,
-0.05118335783481598,
0.005924657918512821,
-0.04527122527360916,
-0.021172842010855675,
0.006223042029887438,
-0.026330644264817238,
0.07170481979846954,
0.039625708013772964,
-0.012840766459703445,
-0.07313792407512665,
0.03444959595799446,
-0.0... | 0.158223 |
delegate to individual nodes for IPAM. - list - ``["fd00::/104"]`` \* - :spelling:ignore:`ipam.operator.externalAPILimitBurstSize` - The maximum burst size when rate limiting access to external APIs. Also known as the token bucket capacity. - int - ``20`` \* - :spelling:ignore:`ipam.operator.externalAPILimitQPS` - The maximum queries per second when rate limiting access to external APIs. Also known as the bucket refill rate, which is used to refill the bucket up to the burst size capacity. - float - ``4.0`` \* - :spelling:ignore:`iptablesRandomFully` - Configure iptables--random-fully. Disabled by default. View https://github.com/cilium/cilium/issues/13037 for more information. - bool - ``false`` \* - :spelling:ignore:`ipv4.enabled` - Enable IPv4 support. - bool - ``true`` \* - :spelling:ignore:`ipv4NativeRoutingCIDR` - Allows to explicitly specify the IPv4 CIDR for native routing. When specified, Cilium assumes networking for this CIDR is preconfigured and hands traffic destined for that range to the Linux network stack without applying any SNAT. Generally speaking, specifying a native routing CIDR implies that Cilium can depend on the underlying networking stack to route packets to their destination. To offer a concrete example, if Cilium is configured to use direct routing and the Kubernetes CIDR is included in the native routing CIDR, the user must configure the routes to reach pods, either manually or by setting the auto-direct-node-routes flag. - string - ``""`` \* - :spelling:ignore:`ipv6.enabled` - Enable IPv6 support. - bool - ``false`` \* - :spelling:ignore:`ipv6NativeRoutingCIDR` - Allows to explicitly specify the IPv6 CIDR for native routing. When specified, Cilium assumes networking for this CIDR is preconfigured and hands traffic destined for that range to the Linux network stack without applying any SNAT. Generally speaking, specifying a native routing CIDR implies that Cilium can depend on the underlying networking stack to route packets to their destination. To offer a concrete example, if Cilium is configured to use direct routing and the Kubernetes CIDR is included in the native routing CIDR, the user must configure the routes to reach pods, either manually or by setting the auto-direct-node-routes flag. - string - ``""`` \* - :spelling:ignore:`k8s` - Configure Kubernetes specific configuration - object - ``{"requireIPv4PodCIDR":false,"requireIPv6PodCIDR":false}`` \* - :spelling:ignore:`k8s.requireIPv4PodCIDR` - requireIPv4PodCIDR enables waiting for Kubernetes to provide the PodCIDR range via the Kubernetes node resource - bool - ``false`` \* - :spelling:ignore:`k8s.requireIPv6PodCIDR` - requireIPv6PodCIDR enables waiting for Kubernetes to provide the PodCIDR range via the Kubernetes node resource - bool - ``false`` \* - :spelling:ignore:`k8sClientExponentialBackoff` - Configure exponential backoff for client-go in Cilium agent. - object - ``{"backoffBaseSeconds":1,"backoffMaxDurationSeconds":120,"enabled":true}`` \* - :spelling:ignore:`k8sClientExponentialBackoff.backoffBaseSeconds` - Configure base (in seconds) for exponential backoff. - int - ``1`` \* - :spelling:ignore:`k8sClientExponentialBackoff.backoffMaxDurationSeconds` - Configure maximum duration (in seconds) for exponential backoff. - int - ``120`` \* - :spelling:ignore:`k8sClientExponentialBackoff.enabled` - Enable exponential backoff for client-go in Cilium agent. - bool - ``true`` \* - :spelling:ignore:`k8sClientRateLimit` - Configure the client side rate limit for the agent If the amount of requests to the Kubernetes API server exceeds the configured rate limit, the agent will start to throttle requests by delaying them until there is budget or the request times out. - object - ``{"burst":null,"operator":{"burst":null,"qps":null},"qps":null}`` \* - :spelling:ignore:`k8sClientRateLimit.burst` - The burst request rate in requests per second. The rate limiter will allow short bursts with a higher rate. - int - 20 \* - :spelling:ignore:`k8sClientRateLimit.operator` - Configure the client side rate limit for the Cilium Operator - object - ``{"burst":null,"qps":null}`` \* - :spelling:ignore:`k8sClientRateLimit.operator.burst` - The burst request rate in requests per second. The rate limiter will allow short bursts with a higher rate. - int - 200 \* - :spelling:ignore:`k8sClientRateLimit.operator.qps` - The sustained request rate in requests per second. - int - 100 \* - :spelling:ignore:`k8sClientRateLimit.qps` - The sustained request rate | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.008091166615486145,
0.00808901246637106,
-0.06970519572496414,
-0.0037088608369231224,
-0.040247414261102676,
-0.08303907513618469,
0.003301294520497322,
0.016939694061875343,
-0.01572054997086525,
0.04591422155499458,
0.06755230575799942,
-0.06502842903137207,
0.03210700303316116,
-0.05... | 0.122166 |
- :spelling:ignore:`k8sClientRateLimit.operator.burst` - The burst request rate in requests per second. The rate limiter will allow short bursts with a higher rate. - int - 200 \* - :spelling:ignore:`k8sClientRateLimit.operator.qps` - The sustained request rate in requests per second. - int - 100 \* - :spelling:ignore:`k8sClientRateLimit.qps` - The sustained request rate in requests per second. - int - 10 \* - :spelling:ignore:`k8sNetworkPolicy.enabled` - Enable support for K8s NetworkPolicy - bool - ``true`` \* - :spelling:ignore:`k8sServiceHost` - Kubernetes service host - use "auto" for automatic lookup from the cluster-info ConfigMap - string - ``""`` \* - :spelling:ignore:`k8sServiceHostRef` - Configure the Kubernetes service endpoint dynamically using a ConfigMap. Mutually exclusive with ``k8sServiceHost``. - object - ``{"key":null,"name":null}`` \* - :spelling:ignore:`k8sServiceHostRef.key` - Key in the ConfigMap containing the Kubernetes service endpoint - string - ``nil`` \* - :spelling:ignore:`k8sServiceHostRef.name` - name of the ConfigMap containing the Kubernetes service endpoint - string - ``nil`` \* - :spelling:ignore:`k8sServiceLookupConfigMapName` - When ``k8sServiceHost=auto``\ , allows to customize the configMap name. It defaults to ``cluster-info``. - string - ``""`` \* - :spelling:ignore:`k8sServiceLookupNamespace` - When ``k8sServiceHost=auto``\ , allows to customize the namespace that contains ``k8sServiceLookupConfigMapName``. It defaults to ``kube-public``. - string - ``""`` \* - :spelling:ignore:`k8sServicePort` - Kubernetes service port - string - ``""`` \* - :spelling:ignore:`keepDeprecatedLabels` - Keep the deprecated selector labels when deploying Cilium DaemonSet. - bool - ``false`` \* - :spelling:ignore:`keepDeprecatedProbes` - Keep the deprecated probes when deploying Cilium DaemonSet - bool - ``false`` \* - :spelling:ignore:`kubeConfigPath` - Kubernetes config path - string - ``"~/.kube/config"`` \* - :spelling:ignore:`kubeProxyReplacement` - Configure the kube-proxy replacement in Cilium BPF datapath Valid options are "true" or "false". ref: https://docs.cilium.io/en/stable/network/kubernetes/kubeproxy-free/ @schema@ type: [string, boolean] @schema@ - string - ``"false"`` \* - :spelling:ignore:`kubeProxyReplacementHealthzBindAddr` - healthz server bind address for the kube-proxy replacement. To enable set the value to '0.0.0.0:10256' for all ipv4 addresses and this '[::]:10256' for all ipv6 addresses. By default it is disabled. - string - ``""`` \* - :spelling:ignore:`l2NeighDiscovery.enabled` - Enable L2 neighbor discovery in the agent - bool - ``false`` \* - :spelling:ignore:`l2announcements` - Configure L2 announcements - object - ``{"enabled":false}`` \* - :spelling:ignore:`l2announcements.enabled` - Enable L2 announcements - bool - ``false`` \* - :spelling:ignore:`l2podAnnouncements` - Configure L2 pod announcements - object - ``{"enabled":false,"interface":"eth0"}`` \* - :spelling:ignore:`l2podAnnouncements.enabled` - Enable L2 pod announcements - bool - ``false`` \* - :spelling:ignore:`l2podAnnouncements.interface` - Interface used for sending Gratuitous ARP pod announcements - string - ``"eth0"`` \* - :spelling:ignore:`l7Proxy` - Enable Layer 7 network policy. - bool - ``true`` \* - :spelling:ignore:`livenessProbe.failureThreshold` - failure threshold of liveness probe - int - ``10`` \* - :spelling:ignore:`livenessProbe.periodSeconds` - interval between checks of the liveness probe - int - ``30`` \* - :spelling:ignore:`livenessProbe.requireK8sConnectivity` - whether to require k8s connectivity as part of the check. - bool - ``false`` \* - :spelling:ignore:`loadBalancer` - Configure service load balancing - object - ``{"acceleration":"disabled","l7":{"algorithm":"round\_robin","backend":"disabled","ports":[]}}`` \* - :spelling:ignore:`loadBalancer.acceleration` - acceleration is the option to accelerate service handling via XDP Applicable values can be: disabled (do not use XDP), native (XDP BPF program is run directly out of the networking driver's early receive path), or best-effort (use native mode XDP acceleration on devices that support it). - string - ``"disabled"`` \* - :spelling:ignore:`loadBalancer.l7` - L7 LoadBalancer - object - ``{"algorithm":"round\_robin","backend":"disabled","ports":[]}`` \* - :spelling:ignore:`loadBalancer.l7.algorithm` - Default LB algorithm The default LB algorithm to be used for services, which can be overridden by the service annotation (e.g. service.cilium.io/lb-l7-algorithm) Applicable values: round\_robin, least\_request, random - string - ``"round\_robin"`` \* - :spelling:ignore:`loadBalancer.l7.backend` - Enable L7 service load balancing via envoy proxy. The request to a k8s service, which has specific annotation e.g. service.cilium.io/lb-l7, will be forwarded to the local backend proxy to be load balanced to the service | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.026515323668718338,
0.010710837319493294,
0.0015498580178245902,
0.009087209589779377,
-0.062192149460315704,
0.013495920225977898,
-0.024891039356589317,
-0.04268045350909233,
0.032249245792627335,
0.04308898746967316,
0.0052801198326051235,
-0.06277620792388916,
-0.006204004865139723,
... | 0.125299 |
(e.g. service.cilium.io/lb-l7-algorithm) Applicable values: round\_robin, least\_request, random - string - ``"round\_robin"`` \* - :spelling:ignore:`loadBalancer.l7.backend` - Enable L7 service load balancing via envoy proxy. The request to a k8s service, which has specific annotation e.g. service.cilium.io/lb-l7, will be forwarded to the local backend proxy to be load balanced to the service endpoints. Please refer to docs for supported annotations for more configuration. Applicable values: - envoy: Enable L7 load balancing via envoy proxy. This will automatically set enable-envoy-config as well. - disabled: Disable L7 load balancing by way of service annotation. - string - ``"disabled"`` \* - :spelling:ignore:`loadBalancer.l7.ports` - List of ports from service to be automatically redirected to above backend. Any service exposing one of these ports will be automatically redirected. Fine-grained control can be achieved by using the service annotation. - list - ``[]`` \* - :spelling:ignore:`localRedirectPolicies.addressMatcherCIDRs` - Limit the allowed addresses in Address Matcher rule of Local Redirect Policies to the given CIDRs. @schema@ type: [null, array] @schema@ - string - ``nil`` \* - :spelling:ignore:`localRedirectPolicies.enabled` - Enable local redirect policies. - bool - ``false`` \* - :spelling:ignore:`localRedirectPolicy` - Enable Local Redirect Policy (deprecated, please use 'localRedirectPolicies.enabled' instead) - bool - ``false`` \* - :spelling:ignore:`logSystemLoad` - Enables periodic logging of system load - bool - ``false`` \* - :spelling:ignore:`maglev` - Configure maglev consistent hashing - object - ``{}`` \* - :spelling:ignore:`monitor` - cilium-monitor sidecar. - object - ``{"enabled":false}`` \* - :spelling:ignore:`monitor.enabled` - Enable the cilium-monitor sidecar. - bool - ``false`` \* - :spelling:ignore:`name` - Agent daemonset name. - string - ``"cilium"`` \* - :spelling:ignore:`namespaceOverride` - namespaceOverride allows to override the destination namespace for Cilium resources. - string - ``""`` \* - :spelling:ignore:`nat.mapStatsEntries` - Number of the top-k SNAT map connections to track in Cilium statedb. - int - ``32`` \* - :spelling:ignore:`nat.mapStatsInterval` - Interval between how often SNAT map is counted for stats. - string - ``"30s"`` \* - :spelling:ignore:`nat46x64Gateway` - Configure standalone NAT46/NAT64 gateway - object - ``{"enabled":false}`` \* - :spelling:ignore:`nat46x64Gateway.enabled` - Enable RFC6052-prefixed translation - bool - ``false`` \* - :spelling:ignore:`nodeIPAM.enabled` - Configure Node IPAM ref: https://docs.cilium.io/en/stable/network/node-ipam/ - bool - ``false`` \* - :spelling:ignore:`nodePort` - Configure N-S k8s service loadbalancing - object - ``{"addresses":null,"autoProtectPortRange":true,"bindProtection":true,"enableHealthCheck":true,"enableHealthCheckLoadBalancerIP":false}`` \* - :spelling:ignore:`nodePort.addresses` - List of CIDRs for choosing which IP addresses assigned to native devices are used for NodePort load-balancing. By default this is empty and the first suitable, preferably private, IPv4 and IPv6 address assigned to each device is used. Example: addresses: ["192.168.1.0/24", "2001::/64"] - string - ``nil`` \* - :spelling:ignore:`nodePort.autoProtectPortRange` - Append NodePort range to ip\_local\_reserved\_ports if clash with ephemeral ports is detected. - bool - ``true`` \* - :spelling:ignore:`nodePort.bindProtection` - Set to true to prevent applications binding to service ports. - bool - ``true`` \* - :spelling:ignore:`nodePort.enableHealthCheck` - Enable healthcheck nodePort server for NodePort services - bool - ``true`` \* - :spelling:ignore:`nodePort.enableHealthCheckLoadBalancerIP` - Enable access of the healthcheck nodePort on the LoadBalancerIP. Needs EnableHealthCheck to be enabled - bool - ``false`` \* - :spelling:ignore:`nodeSelector` - Node selector for cilium-agent. - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`nodeSelectorLabels` - Enable/Disable use of node label based identity - bool - ``false`` \* - :spelling:ignore:`nodeinit.affinity` - Affinity for cilium-nodeinit - object - ``{}`` \* - :spelling:ignore:`nodeinit.annotations` - Annotations to be added to all top-level nodeinit objects (resources under templates/cilium-nodeinit) - object - ``{}`` \* - :spelling:ignore:`nodeinit.bootstrapFile` - bootstrapFile is the location of the file where the bootstrap timestamp is written by the node-init DaemonSet - string - ``"/tmp/cilium-bootstrap.d/cilium-bootstrap-time"`` \* - :spelling:ignore:`nodeinit.enabled` - Enable the node initialization DaemonSet - bool - ``false`` \* - :spelling:ignore:`nodeinit.extraEnv` - Additional nodeinit environment variables. - list - ``[]`` \* - :spelling:ignore:`nodeinit.extraVolumeMounts` - Additional nodeinit volumeMounts. - list - | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.08588110655546188,
-0.0027797420043498278,
-0.03345810994505882,
-0.013924890197813511,
-0.07758287340402603,
-0.04579874873161316,
-0.042143601924180984,
-0.018238235265016556,
-0.044184911996126175,
0.02303808182477951,
-0.08415357023477554,
-0.024034962058067322,
0.025611234828829765,
... | 0.142638 |
file where the bootstrap timestamp is written by the node-init DaemonSet - string - ``"/tmp/cilium-bootstrap.d/cilium-bootstrap-time"`` \* - :spelling:ignore:`nodeinit.enabled` - Enable the node initialization DaemonSet - bool - ``false`` \* - :spelling:ignore:`nodeinit.extraEnv` - Additional nodeinit environment variables. - list - ``[]`` \* - :spelling:ignore:`nodeinit.extraVolumeMounts` - Additional nodeinit volumeMounts. - list - ``[]`` \* - :spelling:ignore:`nodeinit.extraVolumes` - Additional nodeinit volumes. - list - ``[]`` \* - :spelling:ignore:`nodeinit.image` - node-init image. - object - ``{"digest":"sha256:50b9cf9c280096b59b80d2fc8ee6638facef79ac18998a22f0cbc40d5d28c16f","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/startup-script","tag":"1763560095-8f36c34","useDigest":true}`` \* - :spelling:ignore:`nodeinit.nodeSelector` - Node labels for nodeinit pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`nodeinit.podAnnotations` - Annotations to be added to node-init pods. - object - ``{}`` \* - :spelling:ignore:`nodeinit.podLabels` - Labels to be added to node-init pods. - object - ``{}`` \* - :spelling:ignore:`nodeinit.podSecurityContext` - Security Context for cilium-node-init pods. - object - ``{"appArmorProfile":{"type":"Unconfined"}}`` \* - :spelling:ignore:`nodeinit.podSecurityContext.appArmorProfile` - AppArmorProfile options for the ``cilium-node-init`` and init containers - object - ``{"type":"Unconfined"}`` \* - :spelling:ignore:`nodeinit.prestop` - prestop offers way to customize prestop nodeinit script (pre and post position) - object - ``{"postScript":"","preScript":""}`` \* - :spelling:ignore:`nodeinit.priorityClassName` - The priority class to use for the nodeinit pod. - string - ``""`` \* - :spelling:ignore:`nodeinit.resources` - nodeinit resource limits & requests ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ - object - ``{"requests":{"cpu":"100m","memory":"100Mi"}}`` \* - :spelling:ignore:`nodeinit.securityContext` - Security context to be added to nodeinit pods. - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"add":["SYS\_MODULE","NET\_ADMIN","SYS\_ADMIN","SYS\_CHROOT","SYS\_PTRACE"]},"privileged":false,"seLinuxOptions":{"level":"s0","type":"spc\_t"}}`` \* - :spelling:ignore:`nodeinit.startup` - startup offers way to customize startup nodeinit script (pre and post position) - object - ``{"postScript":"","preScript":""}`` \* - :spelling:ignore:`nodeinit.tolerations` - Node tolerations for nodeinit scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[{"operator":"Exists"}]`` \* - :spelling:ignore:`nodeinit.updateStrategy` - node-init update strategy - object - ``{"type":"RollingUpdate"}`` \* - :spelling:ignore:`nodeinit.waitForCloudInit` - wait for Cloud init to finish on the host and assume the node has cloud init installed - bool - ``false`` \* - :spelling:ignore:`operator.affinity` - Affinity for cilium-operator - object - ``{"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"io.cilium/app":"operator"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`operator.annotations` - Annotations to be added to all top-level cilium-operator objects (resources under templates/cilium-operator) - object - ``{}`` \* - :spelling:ignore:`operator.dashboards` - Grafana dashboards for cilium-operator grafana can import dashboards based on the label and value ref: https://github.com/grafana/helm-charts/tree/main/charts/grafana#sidecar-for-dashboards - object - ``{"annotations":{},"enabled":false,"label":"grafana\_dashboard","labelValue":"1","namespace":null}`` \* - :spelling:ignore:`operator.dnsPolicy` - DNS policy for Cilium operator pods. Ref: https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy - string - ``""`` \* - :spelling:ignore:`operator.enabled` - Enable the cilium-operator component (required). - bool - ``true`` \* - :spelling:ignore:`operator.endpointGCInterval` - Interval for endpoint garbage collection. - string - ``"5m0s"`` \* - :spelling:ignore:`operator.extraArgs` - Additional cilium-operator container arguments. - list - ``[]`` \* - :spelling:ignore:`operator.extraEnv` - Additional cilium-operator environment variables. - list - ``[]`` \* - :spelling:ignore:`operator.extraHostPathMounts` - Additional cilium-operator hostPath mounts. - list - ``[]`` \* - :spelling:ignore:`operator.extraVolumeMounts` - Additional cilium-operator volumeMounts. - list - ``[]`` \* - :spelling:ignore:`operator.extraVolumes` - Additional cilium-operator volumes. - list - ``[]`` \* - :spelling:ignore:`operator.hostNetwork` - HostNetwork setting - bool - ``true`` \* - :spelling:ignore:`operator.hostUsers` - HostUsers setting (must be true if hostNetwork is true) - bool - ``true`` \* - :spelling:ignore:`operator.identityGCInterval` - Interval for identity garbage collection. - string - ``"15m0s"`` \* - :spelling:ignore:`operator.identityHeartbeatTimeout` - Timeout for identity heartbeats. - string - ``"30m0s"`` \* - :spelling:ignore:`operator.image` - cilium-operator image. - object - ``{"alibabacloudDigest":"","awsDigest":"","azureDigest":"","genericDigest":"","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/operator","suffix":"-ci","tag":"latest","useDigest":false}`` \* - :spelling:ignore:`operator.nodeGCInterval` - Interval for cilium node garbage collection. - string - ``"5m0s"`` \* - :spelling:ignore:`operator.nodeSelector` - Node labels for cilium-operator pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`operator.podAnnotations` - Annotations to be added to cilium-operator pods - object - ``{}`` \* - :spelling:ignore:`operator.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`operator.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`operator.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.024168284609913826,
0.05842440947890282,
-0.037310726940631866,
-0.015924343839287758,
0.11912707984447479,
-0.07020031660795212,
-0.038473840802907944,
0.06894511729478836,
0.01206669956445694,
0.004482110496610403,
0.05548018589615822,
-0.04984017461538315,
0.013063095510005951,
-0.05... | 0.102667 |
- ``{}`` \* - :spelling:ignore:`operator.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`operator.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`operator.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by ``maxUnavailable: null`` - string - ``nil`` \* - :spelling:ignore:`operator.podDisruptionBudget.unhealthyPodEvictionPolicy` - How are unhealthy, but running, pods counted for eviction - string - ``nil`` \* - :spelling:ignore:`operator.podLabels` - Labels to be added to cilium-operator pods - object - ``{}`` \* - :spelling:ignore:`operator.podSecurityContext` - Security context to be added to cilium-operator pods - object - ``{"seccompProfile":{"type":"RuntimeDefault"}}`` \* - :spelling:ignore:`operator.pprof.address` - Configure pprof listen address for cilium-operator - string - ``"localhost"`` \* - :spelling:ignore:`operator.pprof.blockProfileRate` - Enable goroutine blocking profiling for cilium-operator and set the rate of sampled events in nanoseconds (set to 1 to sample all events [warning: performance overhead]) - int - ``0`` \* - :spelling:ignore:`operator.pprof.enabled` - Enable pprof for cilium-operator - bool - ``false`` \* - :spelling:ignore:`operator.pprof.mutexProfileFraction` - Enable mutex contention profiling for cilium-operator and set the fraction of sampled events (set to 1 to sample all events) - int - ``0`` \* - :spelling:ignore:`operator.pprof.port` - Configure pprof listen port for cilium-operator - int - ``6061`` \* - :spelling:ignore:`operator.priorityClassName` - The priority class to use for cilium-operator - string - ``""`` \* - :spelling:ignore:`operator.prometheus` - Enable prometheus metrics for cilium-operator on the configured port at /metrics - object - ``{"enabled":true,"metricsService":false,"port":9963,"serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","jobLabel":"","labels":{},"metricRelabelings":null,"relabelings":null,"scrapeTimeout":null},"tls":{"enabled":false,"server":{"existingSecret":"","mtls":{"enabled":false}}}}`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.annotations` - Annotations to add to ServiceMonitor cilium-operator - object - ``{}`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.enabled` - Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.interval` - Interval for scrape metrics. - string - ``"10s"`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.jobLabel` - jobLabel to add for ServiceMonitor cilium-operator - string - ``""`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.labels` - Labels to add to ServiceMonitor cilium-operator - object - ``{}`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor cilium-operator - string - ``nil`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor cilium-operator - string - ``nil`` \* - :spelling:ignore:`operator.prometheus.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`operator.prometheus.tls` - TLS configuration for Prometheus - object - ``{"enabled":false,"server":{"existingSecret":"","mtls":{"enabled":false}}}`` \* - :spelling:ignore:`operator.prometheus.tls.server.existingSecret` - Name of the Secret containing the certificate, key and CA files for the Prometheus server. - string - ``""`` \* - :spelling:ignore:`operator.removeNodeTaints` - Remove Cilium node taint from Kubernetes nodes that have a healthy Cilium pod running. - bool - ``true`` \* - :spelling:ignore:`operator.replicas` - Number of replicas to run for the cilium-operator deployment - int - ``2`` \* - :spelling:ignore:`operator.resources` - cilium-operator resource limits & requests ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ - object - ``{}`` \* - :spelling:ignore:`operator.rollOutPods` - Roll out cilium-operator pods automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`operator.securityContext` - Security context to be added to cilium-operator pods - object - ``{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}}`` \* - :spelling:ignore:`operator.setNodeNetworkStatus` - Set Node condition NetworkUnavailable to 'false' with the reason 'CiliumIsUp' for nodes that have a healthy Cilium pod. - bool - ``true`` \* - :spelling:ignore:`operator.setNodeTaints` - Taint nodes where Cilium is scheduled but not running. This prevents pods from being scheduled to nodes where Cilium is not the default CNI provider. - string - same as removeNodeTaints \* - :spelling:ignore:`operator.skipCRDCreation` - Skip CRDs creation for cilium-operator - bool - ``false`` \* - :spelling:ignore:`operator.tolerations` - Node tolerations for cilium-operator scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ Toleration for agentNotReadyTaintKey taint is always added to cilium-operator pods. @schema type: [null, array] @schema - list - ``[{"key":"node-role.kubernetes.io/control-plane","operator":"Exists"},{"key":"node-role.kubernetes.io/master","operator":"Exists"},{"key":"node.kubernetes.io/not-ready","operator":"Exists"},{"key":"node.cloudprovider.kubernetes.io/uninitialized","operator":"Exists"}]`` \* - :spelling:ignore:`operator.topologySpreadConstraints` - Pod topology spread constraints for | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.05117471516132355,
0.0007032703724689782,
0.019332850351929665,
0.04168153926730156,
-0.04508112370967865,
-0.026580553501844406,
0.005124459508806467,
-0.027285369113087654,
0.04520579054951668,
-0.0005744000663980842,
0.05305564031004906,
-0.04091296344995499,
0.011809341609477997,
-0.... | 0.153201 |
creation for cilium-operator - bool - ``false`` \* - :spelling:ignore:`operator.tolerations` - Node tolerations for cilium-operator scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ Toleration for agentNotReadyTaintKey taint is always added to cilium-operator pods. @schema type: [null, array] @schema - list - ``[{"key":"node-role.kubernetes.io/control-plane","operator":"Exists"},{"key":"node-role.kubernetes.io/master","operator":"Exists"},{"key":"node.kubernetes.io/not-ready","operator":"Exists"},{"key":"node.cloudprovider.kubernetes.io/uninitialized","operator":"Exists"}]`` \* - :spelling:ignore:`operator.topologySpreadConstraints` - Pod topology spread constraints for cilium-operator - list - ``[]`` \* - :spelling:ignore:`operator.unmanagedPodWatcher.intervalSeconds` - Interval, in seconds, to check if there are any pods that are not managed by Cilium. - int - ``15`` \* - :spelling:ignore:`operator.unmanagedPodWatcher.restart` - Restart any pod that are not managed by Cilium. - bool - ``true`` \* - :spelling:ignore:`operator.unmanagedPodWatcher.selector` - Selector for pods that should be restarted when not managed by Cilium. If not set, defaults to built-in selector "k8s-app=kube-dns". Set to empty string to select all pods. @schema type: [null, string] @schema - string - ``nil`` \* - :spelling:ignore:`operator.updateStrategy` - cilium-operator update strategy - object - ``{"rollingUpdate":{"maxSurge":"25%","maxUnavailable":"50%"},"type":"RollingUpdate"}`` \* - :spelling:ignore:`pmtuDiscovery.enabled` - Enable path MTU discovery to send ICMP fragmentation-needed replies to the client. - bool - ``false`` \* - :spelling:ignore:`pmtuDiscovery.packetizationLayerPMTUD` - Enable kernel probing path MTU discovery for Pods which uses different message sizes to search for correct MTU value. - object - ``{"enabled":true}`` \* - :spelling:ignore:`podAnnotations` - Annotations to be added to agent pods - object - ``{}`` \* - :spelling:ignore:`podLabels` - Labels to be added to agent pods - object - ``{}`` \* - :spelling:ignore:`podSecurityContext` - Security Context for cilium-agent pods. - object - ``{"appArmorProfile":{"type":"Unconfined"},"seccompProfile":{"type":"Unconfined"}}`` \* - :spelling:ignore:`podSecurityContext.appArmorProfile` - AppArmorProfile options for the ``cilium-agent`` and init containers - object - ``{"type":"Unconfined"}`` \* - :spelling:ignore:`policyCIDRMatchMode` - policyCIDRMatchMode is a list of entities that may be selected by CIDR selector. The possible value is "nodes". - string - ``nil`` \* - :spelling:ignore:`policyDenyResponse` - Configure what the response should be to pod egress traffic denied by network policy. Possible values: - none (default) - icmp - string - ``"none"`` \* - :spelling:ignore:`policyEnforcementMode` - The agent can be put into one of the three policy enforcement modes: default, always and never. ref: https://docs.cilium.io/en/stable/security/policy/intro/#policy-enforcement-modes - string - ``"default"`` \* - :spelling:ignore:`pprof.address` - Configure pprof listen address for cilium-agent - string - ``"localhost"`` \* - :spelling:ignore:`pprof.blockProfileRate` - Enable goroutine blocking profiling for cilium-agent and set the rate of sampled events in nanoseconds (set to 1 to sample all events [warning: performance overhead]) - int - ``0`` \* - :spelling:ignore:`pprof.enabled` - Enable pprof for cilium-agent - bool - ``false`` \* - :spelling:ignore:`pprof.mutexProfileFraction` - Enable mutex contention profiling for cilium-agent and set the fraction of sampled events (set to 1 to sample all events) - int - ``0`` \* - :spelling:ignore:`pprof.port` - Configure pprof listen port for cilium-agent - int - ``6060`` \* - :spelling:ignore:`preflight.affinity` - Affinity for cilium-preflight - object - ``{"podAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"k8s-app":"cilium"}},"topologyKey":"kubernetes.io/hostname"}]}}`` \* - :spelling:ignore:`preflight.annotations` - Annotations to be added to all top-level preflight objects (resources under templates/cilium-preflight) - object - ``{}`` \* - :spelling:ignore:`preflight.enabled` - Enable Cilium pre-flight resources (required for upgrade) - bool - ``false`` \* - :spelling:ignore:`preflight.envoy.image` - Envoy pre-flight image. - object - ``{"digest":"sha256:c0d2fbd63c3576d0c397e6c84fe149056e2bbb940d538ab3baebb86f703642d3","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/cilium-envoy","tag":"v1.36.5-1768567475-cb4965cbd69f290aef785c42dfadf6159adca553","useDigest":true}`` \* - :spelling:ignore:`preflight.extraEnv` - Additional preflight environment variables. - list - ``[]`` \* - :spelling:ignore:`preflight.extraVolumeMounts` - Additional preflight volumeMounts. - list - ``[]`` \* - :spelling:ignore:`preflight.extraVolumes` - Additional preflight volumes. - list - ``[]`` \* - :spelling:ignore:`preflight.image` - Cilium pre-flight image. - object - ``{"digest":"","override":null,"pullPolicy":"Always","repository":"quay.io/cilium/cilium-ci","tag":"latest","useDigest":false}`` \* - :spelling:ignore:`preflight.nodeSelector` - Node labels for preflight pod assignment ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`preflight.podAnnotations` - Annotations to be added to preflight pods - object - ``{}`` \* - :spelling:ignore:`preflight.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`preflight.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.031699929386377335,
0.00858449749648571,
0.008618079125881195,
0.026938648894429207,
-0.029277993366122246,
-0.02379065752029419,
0.023243684321641922,
-0.0795641615986824,
0.054342713207006454,
0.0016078223707154393,
0.008266406133770943,
-0.10383281856775284,
0.007760726381093264,
0.00... | 0.179905 |
https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`preflight.podAnnotations` - Annotations to be added to preflight pods - object - ``{}`` \* - :spelling:ignore:`preflight.podDisruptionBudget.enabled` - enable PodDisruptionBudget ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions/ - bool - ``false`` \* - :spelling:ignore:`preflight.podDisruptionBudget.maxUnavailable` - Maximum number/percentage of pods that may be made unavailable - int - ``1`` \* - :spelling:ignore:`preflight.podDisruptionBudget.minAvailable` - Minimum number/percentage of pods that should remain scheduled. When it's set, maxUnavailable must be disabled by ``maxUnavailable: null`` - string - ``nil`` \* - :spelling:ignore:`preflight.podDisruptionBudget.unhealthyPodEvictionPolicy` - How are unhealthy, but running, pods counted for eviction - string - ``nil`` \* - :spelling:ignore:`preflight.podLabels` - Labels to be added to the preflight pod. - object - ``{}`` \* - :spelling:ignore:`preflight.podSecurityContext` - Security context to be added to preflight pods. - object - ``{}`` \* - :spelling:ignore:`preflight.priorityClassName` - The priority class to use for the preflight pod. - string - ``""`` \* - :spelling:ignore:`preflight.readinessProbe.initialDelaySeconds` - For how long kubelet should wait before performing the first probe - int - ``5`` \* - :spelling:ignore:`preflight.readinessProbe.periodSeconds` - interval between checks of the readiness probe - int - ``5`` \* - :spelling:ignore:`preflight.resources` - preflight resource limits & requests ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ - object - ``{}`` \* - :spelling:ignore:`preflight.securityContext` - Security context to be added to preflight pods - object - ``{"allowPrivilegeEscalation":false}`` \* - :spelling:ignore:`preflight.terminationGracePeriodSeconds` - Configure termination grace period for preflight Deployment and DaemonSet. - int - ``1`` \* - :spelling:ignore:`preflight.tofqdnsPreCache` - Path to write the ``--tofqdns-pre-cache`` file to. - string - ``""`` \* - :spelling:ignore:`preflight.tolerations` - Node tolerations for preflight scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[{"operator":"Exists"}]`` \* - :spelling:ignore:`preflight.updateStrategy` - preflight update strategy - object - ``{"type":"RollingUpdate"}`` \* - :spelling:ignore:`preflight.validateCNPs` - By default we should always validate the installed CNPs before upgrading Cilium. This will make sure the user will have the policies deployed in the cluster with the right schema. - bool - ``true`` \* - :spelling:ignore:`priorityClassName` - The priority class to use for cilium-agent. - string - ``""`` \* - :spelling:ignore:`prometheus` - Configure prometheus metrics on the configured port at /metrics - object - ``{"controllerGroupMetrics":["write-cni-file","sync-host-ips","sync-lb-maps-with-k8s-services"],"enabled":false,"metrics":null,"metricsService":false,"port":9962,"serviceMonitor":{"annotations":{},"enabled":false,"interval":"10s","jobLabel":"","labels":{},"metricRelabelings":null,"relabelings":[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}],"scrapeTimeout":null,"trustCRDsExist":false}}`` \* - :spelling:ignore:`prometheus.controllerGroupMetrics` - - Enable controller group metrics for monitoring specific Cilium subsystems. The list is a list of controller group names. The special values of "all" and "none" are supported. The set of controller group names is not guaranteed to be stable between Cilium versions. - list - ``["write-cni-file","sync-host-ips","sync-lb-maps-with-k8s-services"]`` \* - :spelling:ignore:`prometheus.metrics` - Metrics that should be enabled or disabled from the default metric list. The list is expected to be separated by a space. (+metric\_foo to enable metric\_foo , -metric\_bar to disable metric\_bar). ref: https://docs.cilium.io/en/stable/observability/metrics/ - string - ``nil`` \* - :spelling:ignore:`prometheus.serviceMonitor.annotations` - Annotations to add to ServiceMonitor cilium-agent - object - ``{}`` \* - :spelling:ignore:`prometheus.serviceMonitor.enabled` - Enable service monitors. This requires the prometheus CRDs to be available (see https://github.com/prometheus-operator/prometheus-operator/blob/main/example/prometheus-operator-crd/monitoring.coreos.com\_servicemonitors.yaml) - bool - ``false`` \* - :spelling:ignore:`prometheus.serviceMonitor.interval` - Interval for scrape metrics. - string - ``"10s"`` \* - :spelling:ignore:`prometheus.serviceMonitor.jobLabel` - jobLabel to add for ServiceMonitor cilium-agent - string - ``""`` \* - :spelling:ignore:`prometheus.serviceMonitor.labels` - Labels to add to ServiceMonitor cilium-agent - object - ``{}`` \* - :spelling:ignore:`prometheus.serviceMonitor.metricRelabelings` - Metrics relabeling configs for the ServiceMonitor cilium-agent - string - ``nil`` \* - :spelling:ignore:`prometheus.serviceMonitor.relabelings` - Relabeling configs for the ServiceMonitor cilium-agent - list - ``[{"action":"replace","replacement":"${1}","sourceLabels":["\_\_meta\_kubernetes\_pod\_node\_name"],"targetLabel":"node"}]`` \* - :spelling:ignore:`prometheus.serviceMonitor.scrapeTimeout` - Timeout after which scrape is considered to be failed. - string - ``nil`` \* - :spelling:ignore:`prometheus.serviceMonitor.trustCRDsExist` - Set to ``true`` and helm will not check for monitoring.coreos.com/v1 CRDs before deploying - bool - ``false`` \* - :spelling:ignore:`rbac.create` - Enable creation of Resource-Based Access Control configuration. - bool - ``true`` \* - :spelling:ignore:`readinessProbe.failureThreshold` - failure threshold of readiness probe - int - ``3`` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.07726531475782394,
0.003298769937828183,
0.039249908179044724,
0.048802781850099564,
-0.0078132264316082,
0.012371488846838474,
-0.033731911331415176,
-0.05502898991107941,
0.025048717856407166,
0.06888961046934128,
-0.0074315909296274185,
-0.012082894332706928,
-0.03549106419086456,
0.0... | 0.111666 |
``nil`` \* - :spelling:ignore:`prometheus.serviceMonitor.trustCRDsExist` - Set to ``true`` and helm will not check for monitoring.coreos.com/v1 CRDs before deploying - bool - ``false`` \* - :spelling:ignore:`rbac.create` - Enable creation of Resource-Based Access Control configuration. - bool - ``true`` \* - :spelling:ignore:`readinessProbe.failureThreshold` - failure threshold of readiness probe - int - ``3`` \* - :spelling:ignore:`readinessProbe.periodSeconds` - interval between checks of the readiness probe - int - ``30`` \* - :spelling:ignore:`resourceQuotas` - Enable resource quotas for priority classes used in the cluster. - object - ``{"cilium":{"hard":{"pods":"10k"}},"enabled":false,"operator":{"hard":{"pods":"15"}}}`` \* - :spelling:ignore:`resources` - Agent resource limits & requests ref: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ - object - ``{}`` \* - :spelling:ignore:`rollOutCiliumPods` - Roll out cilium agent pods automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`routingMode` - Enable native-routing mode or tunneling mode. Possible values: - "" - native - tunnel - string - ``"tunnel"`` \* - :spelling:ignore:`scheduling` - Scheduling configurations for cilium pods - object - ``{"mode":"anti-affinity"}`` \* - :spelling:ignore:`scheduling.mode` - Mode specifies how Cilium daemonset pods should be scheduled to Nodes. ``anti-affinity`` mode applies a pod anti-affinity rule to the cilium daemonset. Pod anti-affinity may significantly impact scheduling throughput for large clusters. See: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#inter-pod-affinity-and-anti-affinity ``kube-scheduler`` mode forgoes the anti-affinity rule for full scheduling throughput. Kube-scheduler avoids host port conflict when scheduling pods. - string - Defaults to apply a pod anti-affinity rule to the agent pod - ``anti-affinity`` \* - :spelling:ignore:`sctp` - SCTP Configuration Values - object - ``{"enabled":false}`` \* - :spelling:ignore:`sctp.enabled` - Enable SCTP support. NOTE: Currently, SCTP support does not support rewriting ports or multihoming. - bool - ``false`` \* - :spelling:ignore:`secretsNamespaceAnnotations` - Annotations to be added to all cilium-secret namespaces (resources under templates/cilium-secrets-namespace) - object - ``{}`` \* - :spelling:ignore:`secretsNamespaceLabels` - Labels to be added to all cilium-secret namespaces (resources under templates/cilium-secrets-namespace) - object - ``{}`` \* - :spelling:ignore:`securityContext.allowPrivilegeEscalation` - disable privilege escalation - bool - ``false`` \* - :spelling:ignore:`securityContext.capabilities.applySysctlOverwrites` - capabilities for the ``apply-sysctl-overwrites`` init container - list - ``["SYS\_ADMIN","SYS\_CHROOT","SYS\_PTRACE"]`` \* - :spelling:ignore:`securityContext.capabilities.ciliumAgent` - Capabilities for the ``cilium-agent`` container - list - ``["CHOWN","KILL","NET\_ADMIN","NET\_RAW","IPC\_LOCK","SYS\_MODULE","SYS\_ADMIN","SYS\_RESOURCE","DAC\_OVERRIDE","FOWNER","SETGID","SETUID","SYSLOG"]`` \* - :spelling:ignore:`securityContext.capabilities.cleanCiliumState` - Capabilities for the ``clean-cilium-state`` init container - list - ``["NET\_ADMIN","SYS\_MODULE","SYS\_ADMIN","SYS\_RESOURCE"]`` \* - :spelling:ignore:`securityContext.capabilities.mountCgroup` - Capabilities for the ``mount-cgroup`` init container - list - ``["SYS\_ADMIN","SYS\_CHROOT","SYS\_PTRACE"]`` \* - :spelling:ignore:`securityContext.privileged` - Run the pod with elevated privileges - bool - ``false`` \* - :spelling:ignore:`securityContext.seLinuxOptions` - SELinux options for the ``cilium-agent`` and init containers - object - ``{"level":"s0","type":"spc\_t"}`` \* - :spelling:ignore:`serviceAccounts` - Define serviceAccount names for components. - object - Component's fully qualified name. \* - :spelling:ignore:`serviceAccounts.clustermeshcertgen` - Clustermeshcertgen is used if clustermesh.apiserver.tls.auto.method=cronJob - object - ``{"annotations":{},"automount":true,"create":true,"name":"clustermesh-apiserver-generate-certs"}`` \* - :spelling:ignore:`serviceAccounts.corednsMCSAPI` - CorednsMCSAPI is used if clustermesh.mcsapi.corednsAutoConfigure.enabled=true - object - ``{"annotations":{},"automount":true,"create":true,"name":"cilium-coredns-mcsapi-autoconfig"}`` \* - :spelling:ignore:`serviceAccounts.hubblecertgen` - Hubblecertgen is used if hubble.tls.auto.method=cronJob - object - ``{"annotations":{},"automount":true,"create":true,"name":"hubble-generate-certs"}`` \* - :spelling:ignore:`serviceAccounts.nodeinit.enabled` - Enabled is temporary until https://github.com/cilium/cilium-cli/issues/1396 is implemented. Cilium CLI doesn't create the SAs for node-init, thus the workaround. Helm is not affected by this issue. Name and automount can be configured, if enabled is set to true. Otherwise, they are ignored. Enabled can be removed once the issue is fixed. Cilium-nodeinit DS must also be fixed. - bool - ``false`` \* - :spelling:ignore:`serviceNoBackendResponse` - Configure what the response should be to traffic for a service without backends. Possible values: - reject (default) - drop - string - ``"reject"`` \* - :spelling:ignore:`sleepAfterInit` - Do not run Cilium agent when running with clean mode. Useful to completely uninstall Cilium as it will stop Cilium from starting and create artifacts in the node. - bool - ``false`` \* - :spelling:ignore:`socketLB` - Configure socket LB - object - ``{"enabled":false}`` \* - :spelling:ignore:`socketLB.enabled` - Enable socket LB - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy` - Standalone | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.001478018588386476,
0.01545018795877695,
-0.04196251556277275,
0.026545574888586998,
0.05089450255036354,
-0.07260952889919281,
-0.001627715420909226,
-0.009350256994366646,
-0.011496199294924736,
0.06856487691402435,
0.01846577413380146,
-0.09790027886629105,
0.0367007851600647,
0.00499... | 0.174174 |
mode. Useful to completely uninstall Cilium as it will stop Cilium from starting and create artifacts in the node. - bool - ``false`` \* - :spelling:ignore:`socketLB` - Configure socket LB - object - ``{"enabled":false}`` \* - :spelling:ignore:`socketLB.enabled` - Enable socket LB - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy` - Standalone DNS Proxy Configuration Note: The standalone DNS proxy uses the agent's dnsProxy.\* configuration for DNS settings (proxyPort, enableDnsCompression) to ensure consistency. - object - ``{"annotations":{},"automountServiceAccountToken":false,"debug":false,"enabled":false,"image":{"digest":"","override":null,"pullPolicy":"Always","repository":"","tag":"","useDigest":false},"nodeSelector":{"kubernetes.io/os":"linux"},"rollOutPods":false,"serverPort":10095,"tolerations":[],"updateStrategy":{"rollingUpdate":{"maxSurge":2,"maxUnavailable":0},"type":"RollingUpdate"}}`` \* - :spelling:ignore:`standaloneDnsProxy.annotations` - Standalone DNS proxy annotations - object - ``{}`` \* - :spelling:ignore:`standaloneDnsProxy.automountServiceAccountToken` - Standalone DNS proxy auto mount service account token - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy.debug` - Standalone DNS proxy debug mode - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy.enabled` - Enable standalone DNS proxy (alpha feature) - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy.image` - Standalone DNS proxy image - object - ``{"digest":"","override":null,"pullPolicy":"Always","repository":"","tag":"","useDigest":false}`` \* - :spelling:ignore:`standaloneDnsProxy.nodeSelector` - Standalone DNS proxy Node Selector - object - ``{"kubernetes.io/os":"linux"}`` \* - :spelling:ignore:`standaloneDnsProxy.rollOutPods` - Roll out Standalone DNS proxy automatically when configmap is updated. - bool - ``false`` \* - :spelling:ignore:`standaloneDnsProxy.serverPort` - Standalone DNS proxy server port - int - ``10095`` \* - :spelling:ignore:`standaloneDnsProxy.tolerations` - Standalone DNS proxy tolerations - list - ``[]`` \* - :spelling:ignore:`standaloneDnsProxy.updateStrategy` - Standalone DNS proxy update strategy - object - ``{"rollingUpdate":{"maxSurge":2,"maxUnavailable":0},"type":"RollingUpdate"}`` \* - :spelling:ignore:`startupProbe.failureThreshold` - failure threshold of startup probe. Allow Cilium to take up to 600s to start up (300 attempts with 2s between attempts). - int - ``300`` \* - :spelling:ignore:`startupProbe.periodSeconds` - interval between checks of the startup probe - int - ``2`` \* - :spelling:ignore:`synchronizeK8sNodes` - Synchronize Kubernetes nodes to kvstore and perform CNP GC. - bool - ``true`` \* - :spelling:ignore:`sysctlfix` - Configure sysctl override described in #20072. - object - ``{"enabled":true}`` \* - :spelling:ignore:`sysctlfix.enabled` - Enable the sysctl override. When enabled, the init container will mount the /proc of the host so that the ``sysctlfix`` utility can execute. - bool - ``true`` \* - :spelling:ignore:`terminationGracePeriodSeconds` - Configure termination grace period for cilium-agent DaemonSet. - int - ``1`` \* - :spelling:ignore:`tls` - Configure TLS configuration in the agent. - object - ``{"ca":{"cert":"","certValidityDuration":1095,"key":""},"caBundle":{"enabled":false,"key":"ca.crt","name":"cilium-root-ca.crt","useSecret":false},"readSecretsOnlyFromSecretsNamespace":null,"secretSync":{"enabled":null},"secretsBackend":null,"secretsNamespace":{"create":true,"name":"cilium-secrets"}}`` \* - :spelling:ignore:`tls.ca` - Base64 encoded PEM values for the CA certificate and private key. This can be used as common CA to generate certificates used by hubble and clustermesh components. It is neither required nor used when cert-manager is used to generate the certificates. - object - ``{"cert":"","certValidityDuration":1095,"key":""}`` \* - :spelling:ignore:`tls.ca.cert` - Optional CA cert. If it is provided, it will be used by cilium to generate all other certificates. Otherwise, an ephemeral CA is generated. - string - ``""`` \* - :spelling:ignore:`tls.ca.certValidityDuration` - Generated certificates validity duration in days. This will be used for auto generated CA. - int - ``1095`` \* - :spelling:ignore:`tls.ca.key` - Optional CA private key. If it is provided, it will be used by cilium to generate all other certificates. Otherwise, an ephemeral CA is generated. - string - ``""`` \* - :spelling:ignore:`tls.caBundle` - Configure the CA trust bundle used for the validation of the certificates leveraged by hubble and clustermesh. When enabled, it overrides the content of the 'ca.crt' field of the respective certificates, allowing for CA rotation with no down-time. - object - ``{"enabled":false,"key":"ca.crt","name":"cilium-root-ca.crt","useSecret":false}`` \* - :spelling:ignore:`tls.caBundle.enabled` - Enable the use of the CA trust bundle. - bool - ``false`` \* - :spelling:ignore:`tls.caBundle.key` - Entry of the ConfigMap containing the CA trust bundle. - string - ``"ca.crt"`` \* - :spelling:ignore:`tls.caBundle.name` - Name of the ConfigMap containing the CA trust bundle. - string - ``"cilium-root-ca.crt"`` \* - :spelling:ignore:`tls.caBundle.useSecret` - Use a Secret instead of a ConfigMap. - bool - ``false`` \* - :spelling:ignore:`tls.readSecretsOnlyFromSecretsNamespace` - Configure if the | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.07445098459720612,
-0.00857476331293583,
-0.029452331364154816,
-0.02471638284623623,
-0.06954061985015869,
-0.09091231226921082,
-0.08932065218687057,
-0.0429745614528656,
-0.0015296775382012129,
-0.0684199333190918,
0.02227083221077919,
0.025016402825713158,
-0.030460011214017868,
-0.... | 0.103476 |
the ConfigMap containing the CA trust bundle. - string - ``"ca.crt"`` \* - :spelling:ignore:`tls.caBundle.name` - Name of the ConfigMap containing the CA trust bundle. - string - ``"cilium-root-ca.crt"`` \* - :spelling:ignore:`tls.caBundle.useSecret` - Use a Secret instead of a ConfigMap. - bool - ``false`` \* - :spelling:ignore:`tls.readSecretsOnlyFromSecretsNamespace` - Configure if the Cilium Agent will only look in ``tls.secretsNamespace`` for CiliumNetworkPolicy relevant Secrets. If false, the Cilium Agent will be granted READ (GET/LIST/WATCH) access to \*all\* secrets in the entire cluster. This is not recommended and is included for backwards compatibility. This value obsoletes ``tls.secretsBackend``\ , with ``true`` == ``local`` in the old setting, and ``false`` == ``k8s``. - string - ``nil`` \* - :spelling:ignore:`tls.secretSync` - Configures settings for synchronization of TLS Interception Secrets - object - ``{"enabled":null}`` \* - :spelling:ignore:`tls.secretSync.enabled` - Enable synchronization of Secrets for TLS Interception. If disabled and tls.readSecretsOnlyFromSecretsNamespace is set to 'false', then secrets will be read directly by the agent. - string - ``nil`` \* - :spelling:ignore:`tls.secretsBackend` - This configures how the Cilium agent loads the secrets used TLS-aware CiliumNetworkPolicies (namely the secrets referenced by terminatingTLS and originatingTLS). This value is DEPRECATED and will be removed in a future version. Use ``tls.readSecretsOnlyFromSecretsNamespace`` instead. Possible values: - local - k8s - string - ``nil`` \* - :spelling:ignore:`tls.secretsNamespace` - Configures where secrets used in CiliumNetworkPolicies will be looked for - object - ``{"create":true,"name":"cilium-secrets"}`` \* - :spelling:ignore:`tls.secretsNamespace.create` - Create secrets namespace for TLS Interception secrets. - bool - ``true`` \* - :spelling:ignore:`tls.secretsNamespace.name` - Name of TLS Interception secret namespace. - string - ``"cilium-secrets"`` \* - :spelling:ignore:`tmpVolume` - Configure temporary volume for cilium-agent - object - ``{}`` \* - :spelling:ignore:`tolerations` - Node tolerations for agent scheduling to nodes with taints ref: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/ - list - ``[{"operator":"Exists"}]`` \* - :spelling:ignore:`tunnelPort` - Configure VXLAN and Geneve tunnel port. - int - Port 8472 for VXLAN, Port 6081 for Geneve \* - :spelling:ignore:`tunnelProtocol` - Tunneling protocol to use in tunneling mode and for ad-hoc tunnels. Possible values: - "" - vxlan - geneve - string - ``"vxlan"`` \* - :spelling:ignore:`tunnelSourcePortRange` - Configure VXLAN and Geneve tunnel source port range hint. - string - 0-0 to let the kernel driver decide the range \* - :spelling:ignore:`underlayProtocol` - IP family for the underlay. - string - ``"ipv4"`` \* - :spelling:ignore:`updateStrategy` - Cilium agent update strategy - object - ``{"rollingUpdate":{"maxUnavailable":2},"type":"RollingUpdate"}`` \* - :spelling:ignore:`upgradeCompatibility` - upgradeCompatibility helps users upgrading to ensure that the configMap for Cilium will not change critical values to ensure continued operation This flag is not required for new installations. For example: '1.7', '1.8', '1.9' - string - ``nil`` \* - :spelling:ignore:`vtep.cidr` - A space separated list of VTEP device CIDRs, for example "1.1.1.0/24 1.1.2.0/24" - string - ``""`` \* - :spelling:ignore:`vtep.enabled` - Enables VXLAN Tunnel Endpoint (VTEP) Integration (beta) to allow Cilium-managed pods to talk to third party VTEP devices over Cilium tunnel. - bool - ``false`` \* - :spelling:ignore:`vtep.endpoint` - A space separated list of VTEP device endpoint IPs, for example "1.1.1.1 1.1.2.1" - string - ``""`` \* - :spelling:ignore:`vtep.mac` - A space separated list of VTEP device MAC addresses (VTEP MAC), for example "x:x:x:x:x:x y:y:y:y:y:y:y" - string - ``""`` \* - :spelling:ignore:`vtep.mask` - VTEP CIDRs Mask that applies to all VTEP CIDRs, for example "255.255.255.0" - string - ``""`` \* - :spelling:ignore:`waitForKubeProxy` - Wait for KUBE-PROXY-CANARY iptables rule to appear in "wait-for-kube-proxy" init container before launching cilium-agent. More context can be found in the commit message of below PR https://github.com/cilium/cilium/pull/20123 - bool - ``false`` \* - :spelling:ignore:`wellKnownIdentities.enabled` - Enable the use of well-known identities. - bool - ``false`` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
-0.07466314733028412,
0.03687496855854988,
-0.09061045199632645,
0.025249192491173744,
-0.04786495864391327,
-0.06880690902471542,
0.021250607445836067,
0.003301416290923953,
0.046277791261672974,
-0.041796281933784485,
0.06440217792987823,
-0.10104876756668091,
0.07400823384523392,
-0.002... | 0.084529 |
"wait-for-kube-proxy" init container before launching cilium-agent. More context can be found in the commit message of below PR https://github.com/cilium/cilium/pull/20123 - bool - ``false`` \* - :spelling:ignore:`wellKnownIdentities.enabled` - Enable the use of well-known identities. - bool - ``false`` | https://github.com/cilium/cilium/blob/main//Documentation/helm-values.rst | main | cilium | [
0.009348999708890915,
0.03675068914890289,
-0.06360243260860443,
-0.01954496279358864,
-0.012636533007025719,
-0.032530345022678375,
-0.005535622593015432,
-0.06885125488042831,
0.04217080399394035,
-0.0007013934664428234,
0.05892734229564667,
-0.08743413537740707,
-0.003693804144859314,
-... | 0.158012 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io Welcome to Cilium's documentation! ================================== The documentation is divided into the following sections: \* :ref:`k8s\_install\_quick`: Provides a simple tutorial for running a small Cilium setup on your laptop. Intended as an easy way to get your hands dirty applying Cilium security policies between containers. \* :ref:`getting\_started` : Details instructions for installing, configuring, and troubleshooting Cilium in different deployment modes. \* :ref:`network\_policy` : Detailed walkthrough of the policy language structure and the supported formats. \* :ref:`observability` : Provides instructions on setting up and configuring :ref:`hubble\_intro` and configuring :ref:`metrics collection from Cilium and Hubble`. \* :ref:`admin\_guide` : Describes how to troubleshoot Cilium in different deployment modes. \* :ref:`bpf\_guide` : Provides a technical deep dive of eBPF and XDP technology, primarily focused at developers. \* :ref:`api\_ref` : Details the Cilium agent API for interacting with a local Cilium instance. \* :ref:`dev\_guide` : Gives background to those looking to develop and contribute modifications to the Cilium code or documentation. \* :ref:`security\_root` : Provides a one-page resource of best practices for securing Cilium. A `hands-on tutorial `\_ in a live environment is also available for users looking for a way to quickly get started and experiment with Cilium. .. toctree:: :maxdepth: 2 :caption: Overview overview/intro overview/component-overview .. \_getting\_started: .. toctree:: :maxdepth: 2 :caption: Getting Started gettingstarted/k8s-install-default gettingstarted/demo gettingstarted/terminology gettingstarted/gettinghelp .. toctree:: :maxdepth: 2 :caption: Advanced Installation installation/taints installation/k8s-install-helm installation/k8s-install-migration installation/k8s-toc installation/external-toc .. toctree:: :maxdepth: 2 :caption: Networking network/concepts/index network/kubernetes/index network/bgp-toc network/ebpf/index network/clustermesh/index network/egress-gateway-toc network/servicemesh/index network/vtep network/l2-announcements network/node-ipam network/pod-mac-address network/multicast .. toctree:: :maxdepth: 2 :caption: Security security/index security/network/index security/policy/index security/restrict-pod-access security/threat-model .. toctree:: :maxdepth: 2 :caption: Observability :name: observability observability/hubble/index observability/grafana observability/metrics observability/visibility .. toctree:: :maxdepth: 2 :caption: Operations operations/system\_requirements operations/upgrade configuration/index operations/performance/index operations/troubleshooting .. toctree:: :maxdepth: 2 :caption: Community community/governance community/community community/roadmap .. toctree:: :maxdepth: 2 :caption: Contributor Guide contributing/development/index contributing/release/index contributing/testing/index contributing/docs/index api grpcapi sdpapi internals/index .. toctree:: :maxdepth: 2 :caption: Reference cheatsheet cmdref/index helm-reference kvstore further\_reading glossary .. toctree:: :maxdepth: 2 :caption: Reference Guides reference-guides/bpf/index reference-guides/xfrm/index | https://github.com/cilium/cilium/blob/main//Documentation/index.rst | main | cilium | [
0.015150709077715874,
0.042053189128637314,
-0.07255500555038452,
-0.04996337369084358,
0.04918280243873596,
-0.0686020776629448,
-0.036964595317840576,
0.060627423226833344,
-0.02410278469324112,
-0.043192390352487564,
0.07840275019407272,
-0.07893601059913635,
0.044164273887872696,
-0.01... | 0.212407 |
Key-Value Store ############### Cilium uses an external key-value store to exchange information across multiple Cilium instances: Layout ====== All data is stored under a common key prefix: ===================== ==================== Prefix Description ===================== ==================== ``cilium/`` All keys share this common prefix. ``cilium/state/`` State stored by agents, data is automatically recreated on removal or corruption. ===================== ==================== Cluster Nodes ------------- Every agent will register itself as a node in the kvstore and make the following information available to other agents: - Name - IP addresses of the node - Health checking IP addresses - Allocation range of endpoints on the node ============================================================ ==================== Key Value ============================================================ ==================== ``cilium/state/nodes/v1//`` node.Node\_ ============================================================ ==================== .. \_node.Node: https://pkg.go.dev/github.com/cilium/cilium/pkg/node/types#Node All node keys are attached to a lease owned by the agent of the respective node. Services -------- All Kubernetes services are mirrored into the kvstore by the Cilium operator. This is required to implement multi cluster service discovery. ============================================================= ==================== Key Value ============================================================= ==================== ``cilium/state/services/v1///`` serviceStore.ClusterService\_ ============================================================= ==================== .. \_serviceStore.ClusterService: https://pkg.go.dev/github.com/cilium/cilium/pkg/clustermesh/store#ClusterService Identities ---------- Any time a new endpoint is started on a Cilium node, it will determine whether the labels for the endpoint are unique and allocate an identity for that set of labels. These identities are only meaningful within the local cluster. ============================================================= ==================== Key Value ============================================================= ==================== ``cilium/state/identities/v1/id/`` labels.LabelArray\_ ``cilium/state/identities/v1/value//`` identity.NumericIdentity\_ ============================================================= ==================== .. \_identity.NumericIdentity: https://pkg.go.dev/github.com/cilium/cilium/pkg/identity#NumericIdentity .. \_labels.LabelArray: https://pkg.go.dev/github.com/cilium/cilium/pkg/labels#LabelArray Endpoints --------- All endpoint IPs and corresponding identities are mirrored to the kvstore by the agent on the node where the endpoint is launched, to allow peer nodes to configure egress policies to endpoints backed by these IPs. ============================================================= ==================== Key Value ============================================================= ==================== ``cilium/state/ip/v1//`` identity.IPIdentityPair\_ ============================================================= ==================== .. \_identity.IPIdentityPair: https://pkg.go.dev/github.com/cilium/cilium/pkg/identity#IPIdentityPair CiliumNetworkPolicyNodeStatus ----------------------------- If handover to Kubernetes is enabled, then each ``cilium-agent`` will propagate the state of whether it has realized a given CNP to the key-value store instead of directly writing to ``kube-apiserver``. ``cilium-operator`` will listen for updates to this prefix from the key-value store, and will be the sole updater of statuses for CNPs in the cluster. ================================================================ ==================== Key Value ================================================================ ==================== ``cilium/state/cnpstatuses/v2////`` k8s.CNPNSWithMeta\_ ================================================================ ==================== .. \_k8s.CNPNSWithMeta: https://pkg.go.dev/github.com/cilium/cilium/pkg/k8s#CNPNSWithMeta Heartbeat --------- The heartbeat key is periodically updated by the operator to contain the current time and date. It is used by agents to validate that kvstore updates can be received. ====================== ====================== Key Value ====================== ====================== ``cilium/.heartbeat`` Current time and date ====================== ====================== .. \_kvstore\_leases: Leases ====== With a few exceptions, all keys in the key-value store are owned by a particular agent running on a node. All such keys have a lease attached. The lease is renewed automatically. When the lease expires, the key is removed from the key-value store. This guarantees that keys are removed from the key-value store in the event that an agent dies on a particular and never reappears. The lease lifetime is set to 15 minutes. The exact expiration behavior is dependent on the kvstore implementation but the expiration typically occurs after double the lease lifetime. In addition to regular entry leases, all locks in the key-value store are owned by a particular agent running on the node with a separate "lock lease" attached. The lock lease has a default lifetime of 25 seconds. =============================================================== ================ ======================================== Key Lease Timeout Default expiry =============================================================== ================ ======================================== ``cilium/.initlock//`` LockLeaseTTL\_ 25 seconds ``cilium/.heartbeat`` KVstoreLeaseTTL 15 minutes ``cilium/state/cnpstatuses/v2////`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/identities/v1/id/`` None Garbage collected by ``cilium-operator`` ``cilium/state/identities/v1/value//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/ip/v1//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/nodes/v1//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/services/v1///`` KVstoreLeaseTTL\_ 15 minutes =============================================================== ================ ======================================== .. \_LockLeaseTTL: https://pkg.go.dev/github.com/cilium/cilium/pkg/defaults?tab=doc#LockLeaseTTL .. \_KVstoreLeaseTTL: https://pkg.go.dev/github.com/cilium/cilium/pkg/defaults?tab=doc#KVstoreLeaseTTL Caveats and Limitations ======================= If you manually remove and recreate kvstore state when IPSec transparent encryption is enabled, then | https://github.com/cilium/cilium/blob/main//Documentation/kvstore.rst | main | cilium | [
0.005103504750877619,
-0.002938730176538229,
-0.08761757612228394,
0.004778861999511719,
-0.06336279958486557,
-0.014556794427335262,
-0.01973428577184677,
0.024960825219750404,
0.03722011297941208,
-0.029523417353630066,
0.06743373721837997,
-0.1044369488954544,
0.0689227506518364,
-0.068... | 0.167253 |
``cilium/state/identities/v1/id/`` None Garbage collected by ``cilium-operator`` ``cilium/state/identities/v1/value//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/ip/v1//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/nodes/v1//`` KVstoreLeaseTTL\_ 15 minutes ``cilium/state/services/v1///`` KVstoreLeaseTTL\_ 15 minutes =============================================================== ================ ======================================== .. \_LockLeaseTTL: https://pkg.go.dev/github.com/cilium/cilium/pkg/defaults?tab=doc#LockLeaseTTL .. \_KVstoreLeaseTTL: https://pkg.go.dev/github.com/cilium/cilium/pkg/defaults?tab=doc#KVstoreLeaseTTL Caveats and Limitations ======================= If you manually remove and recreate kvstore state when IPSec transparent encryption is enabled, then that may cause permanent connectivity disruption for pods managed by Cilium. Refer to :ref:`xfrm\_state\_staling\_in\_cilium` for further details. Debugging ========= The contents stored in the kvstore can be queued and manipulate using the ``cilium kvstore`` command. For additional details, see the command reference. Example: .. code-block:: shell-session $ cilium kvstore get --recursive cilium/state/nodes/ cilium/state/nodes/v1/default/runtime1 => {"Name":"runtime1","IPAddresses":[{"AddressType":"InternalIP","IP":"10.0.2.15"}],"IPv4AllocCIDR":{"IP":"10.11.0.0","Mask":"//8AAA=="},"IPv6AllocCIDR":{"IP":"f00d::a0f:0:0:0","Mask":"//////////////////8AAA=="},"IPv4HealthIP":"","IPv6HealthIP":""} | https://github.com/cilium/cilium/blob/main//Documentation/kvstore.rst | main | cilium | [
-0.014151395298540592,
0.09647906571626663,
-0.04605373740196228,
0.01332749705761671,
-0.06524641811847687,
-0.06937471777200699,
0.007808629423379898,
-0.011587051674723625,
0.05341988429427147,
-0.08413326740264893,
0.08872541040182114,
-0.122344009578228,
0.00597894424572587,
0.0025909... | 0.151061 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Command Cheatsheet \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Cilium is controlled via an easy command-line interface. This CLI is a single application that takes subcommands that you can find in the command reference guide. .. code-block:: shell-session $ cilium CLI for interacting with the local Cilium Agent Usage: cilium-dbg [command] Available Commands: bpf Direct access to local eBPF maps cleanup Reset the agent state completion Output shell completion code for bash config Cilium configuration options debuginfo Request available debugging information from agent endpoint Manage endpoints identity Manage security identities kvstore Direct access to the kvstore monitor Monitoring policy Manage security policies prefilter Manage XDP CIDR filters service Manage services & loadbalancers status Display status of daemon version Print version information Flags: --config string config file (default is $HOME/.cilium.yaml) -D, --debug Enable debug messages -H, --host string URI to server-side API Use "cilium [command] --help" for more information about a command. All commands and subcommands have the option ``-h`` that will provide information about the options and arguments that the subcommand has. In case of any error in the command, Cilium CLI will return a non-zero status. Command utilities: ================== JSON Output ----------- All the list commands will return a pretty printed list with the information retrieved from Cilium Daemon. If you need something more detailed you can use JSON output, to get the JSON output you can use the global option ``-o json`` .. code-block:: shell-session $ cilium-dbg endpoint list -o json Moreover, Cilium also provides a `JSONPath `\_ support, so detailed information can be extracted. JSONPath template reference can be found in `Kubernetes documentation `\_ .. code-block:: shell-session $ cilium-dbg endpoint list -o jsonpath='{[\*].id}' 29898 38939 56326 $ cilium-dbg endpoint list -o jsonpath='{range [\*]}{@.id}{"="}{@.status.policy.spec.policy-enabled}{"\n"}{end}' 29898=none 38939=none 56326=none Shell Tab-completion -------------------- If you use bash or zsh, Cilium CLI can provide tab completion for subcommands. If you want to install tab completion, you should run the following command in your terminal. .. code-block:: shell-session $ source <(cilium completion) If you want to have Cilium completion always loaded, you can install using the following: .. code-block:: shell-session $ echo "source <(cilium completion)" >> ~/.bashrc Command examples: ================= Basics ------ Check the status of the agent .. code-block:: shell-session $ cilium-dbg status KVStore: Ok Etcd: 172.17.0.3:4001 ContainerRuntime: Ok Kubernetes: Disabled Cilium: Ok OK NodeMonitor: Listening for events on 2 CPUs with 64x4096 of shared memory Cilium health daemon: Ok Controller Status: 6/6 healthy Proxy Status: OK, ip 10.15.28.238, port-range 10000-20000 Cluster health: 1/1 reachable (2018-04-11T07:33:09Z) $ Get a detailed status of the agent: .. code-block:: shell-session $ cilium-dbg status --all-controllers --all-health --all-redirects KVStore: Ok Etcd: 172.17.0.3:4001 ContainerRuntime: Ok Kubernetes: Disabled Cilium: Ok OK NodeMonitor: Listening for events on 2 CPUs with 64x4096 of shared memory Cilium health daemon: Ok Controller Status: 6/6 healthy Name Last success Last error Count Message kvstore-lease-keepalive 2m52s ago never 0 no error ipcache-bpf-garbage-collection 2m50s ago never 0 no error resolve-identity-29898 2m50s ago never 0 no error sync-identity-to-k8s-pod (29898) 50s ago never 0 no error sync-IPv4-identity-mapping (29898) 2m49s ago never 0 no error sync-IPv6-identity-mapping (29898) 2m49s ago never 0 no error Proxy Status: OK, ip 10.15.28.238, port-range 10000-20000 Cluster health: 1/1 reachable (2018-04-11T07:32:09Z) Name IP Reachable Endpoints reachable runtime (localhost) 10.0.2.15 true false $ Get the current agent configuration .. code-block:: shell-session cilium-dbg config Get list of all imported policy rules .. code-block:: shell-session cilium-dbg policy get Monitoring ~~~~~~~~~~~ Monitor cilium-dbg datapath notifications .. code-block:: shell-session cilium-dbg monitor Verbose output (including debug if enabled) .. code-block:: shell-session | https://github.com/cilium/cilium/blob/main//Documentation/cheatsheet.rst | main | cilium | [
0.011539934203028679,
0.029722638428211212,
-0.1031070277094841,
-0.023683711886405945,
-0.0186128169298172,
-0.02678346261382103,
-0.01280136127024889,
0.07022609561681747,
0.01249983161687851,
-0.0006827163160778582,
0.04488154128193855,
-0.1298273205757141,
0.047274235635995865,
-0.0526... | 0.200173 |
reachable runtime (localhost) 10.0.2.15 true false $ Get the current agent configuration .. code-block:: shell-session cilium-dbg config Get list of all imported policy rules .. code-block:: shell-session cilium-dbg policy get Monitoring ~~~~~~~~~~~ Monitor cilium-dbg datapath notifications .. code-block:: shell-session cilium-dbg monitor Verbose output (including debug if enabled) .. code-block:: shell-session cilium-dbg monitor -v Extra verbose output (including packet dissection) .. code-block:: shell-session cilium-dbg monitor -v -v Filter for only the events related to endpoint .. code-block:: shell-session cilium-dbg monitor --related-to= Filter for only events on layer 7 .. code-block:: shell-session cilium-dbg monitor -t L7 Show notifications only for dropped packet events .. code-block:: shell-session cilium-dbg monitor --type drop Don't dissect packet payload, display payload in hex information .. code-block:: shell-session cilium-dbg monitor -v -v --hex Connectivity ------------ Check cluster Connectivity .. code-block:: shell-session cilium-health status There is also a `blog post `\_ related to this tool. Endpoints --------- Get list of all local endpoints .. code-block:: shell-session cilium-dbg endpoint list Get detailed view of endpoint properties and state .. code-block:: shell-session cilium-dbg endpoint get Show recent endpoint specific log entries .. code-block:: shell-session cilium-dbg endpoint log Enable debugging output on the cilium-dbg monitor for this endpoint .. code-block:: shell-session cilium-dbg endpoint config Debug=true Loadbalancing ------------- Get list of loadbalancer services .. code-block:: shell-session cilium-dbg service list Or you can get the loadbalancer information using bpf list .. code-block:: shell-session cilium-dbg bpf lb list eBPF ---- Checking logs for verifier issue .. code-block:: shell-session journalctl -u cilium-dbg | grep -B20 -F10 Verifier List connection tracking entries: .. code-block:: shell-session sudo cilium-dbg bpf ct list Flush connection tracking entries: .. code-block:: shell-session sudo cilium-dbg bpf ct flush Kubernetes examples: ===================== If you running Cilium on top of Kubernetes you may also want a way to list all cilium endpoints or policies from a single Kubectl commands. Cilium provides all this information to the user by using `Kubernetes Resource Definitions `\_: Policies --------- In Kubernetes you can use two kinds of policies, Kubernetes Network Policies or Cilium Network Policies. Both can be retrieved from the ``kubectl`` command: .. code-block:: shell-session :name: Kubernetes Network Policies :caption: Kubernetes Network Policies kubectl get netpol .. code-block:: shell-session :name: Kubernetes Cilium Policies :caption: Kubernetes Cilium Policies $ kubectl get cnp NAME AGE rule1 3m $ kubectl get cnp rule1 NAME AGE rule1 3m $ kubectl get cnp rule1 -o json Endpoints ---------- To retrieve a list of all endpoints managed by cilium, ``Cilium Endpoint`` resource can be used. .. code-block:: shell-session $ kubectl get cep NAME AGE 34e299f0-b25c2fef 41s 34e299f0-dd86986c 42s 4d088f48-83e4f98d 2m 4d088f48-d04ab55f 2m 5c6211b5-9217a4d1 1m 5c6211b5-dccc3d24 1m 700e0976-6cb50b02 3m 700e0976-afd3a30c 3m 78092a35-4874ed16 1m 78092a35-4b08b92b 1m 9b74f61f-14571299 7s 9b74f61f-f9a96f4a 7s $ kubectl get cep 700e0976-6cb50b02 -o json $ kubectl get cep -o jsonpath='{range .items[\*]}{@.status.id}{"="}{@.status.status.policy.spec.policy-enabled}{"\n"}{end}' 30391=ingress 5766=ingress 51796=none 40355=none | https://github.com/cilium/cilium/blob/main//Documentation/cheatsheet.rst | main | cilium | [
-0.009774752892553806,
0.038433533161878586,
-0.05924814194440842,
0.009452343918383121,
0.034392669796943665,
-0.0541834719479084,
0.02485467679798603,
0.022269388660788536,
0.016616089269518852,
0.029527314007282257,
0.04603506624698639,
-0.1027764230966568,
-0.006088549736887217,
-0.030... | 0.152026 |
.. This file was autogenerated via Documentation/update-codeowners.sh, do not edit manually Code owners are used by the Cilium community to consolidate common knowledge into teams that can provide consistent and actionable feedback to contributors. This section will describe groups of teams and suggestions about the focus areas for review. The primary motivation for these teams is to provide structure around review processes to ensure that contributors know how to reach out to community members to conduct discussions, ensure contributions meet the expectations of the community, and align on the direction of proposed changes. Furthermore, while these teams are primarily drawn upon to provide review on specific pull requests, they are also encouraged to self-organize around how to make improvements to their areas of the Cilium project over time. Any committer may self-nominate to code owner teams. Reach out to the core team on the #committers channel in Slack to coordinate. Committers do not require expert knowledge in an area in order to join a code owner team, only a willingness to engage in discussions and learn about the area. Project-wide ++++++++++++ These code owners may provide feedback for Pull Requests submitted to any repository in the Cilium project: - `@cilium/api `\_\_: Ensure the backwards-compatibility of Cilium REST and gRPC APIs, excluding Hubble which is owned by `@cilium/sig-hubble-api `\_\_. - `@cilium/build `\_\_: Provide feedback on languages and scripting used for build and packaging system: Make, Shell, Docker. - `@cilium/cli `\_\_: Provide user experience feedback on changes to Command-Line Interfaces. These owners are a stand-in for the user community to bring a user perspective to the review process. Consider how information is presented, consistency of flags and options. - `@cilium/committers `\_\_: Cilium contributors with commit access. - `@cilium/ci-structure `\_\_: Provide guidance around the best use of Cilium project continuous integration and testing infrastructure, including GitHub actions, VM helpers, testing frameworks, etc. - `@cilium/community `\_\_: Maintain files that refer to Cilium community users such as USERS.md. - `@cilium/contributing `\_\_: Encourage practices that ensure an inclusive contributor community. Review tooling and scripts used by contributors. - `@cilium/docs-structure `\_\_: Ensure the consistency and layout of documentation. General feedback on the use of Sphinx, how to communicate content clearly to the community. This code owner is not expected to validate the technical correctness of submissions. Correctness is typically handled by another code owner group which is also assigned to any given piece of documentation. - `@cilium/sig-foundations `\_\_: Review changes to the core libraries and provide guidance to overall software architecture. - `@cilium/github-sec `\_\_: Responsible for maintaining the security of repositories in the Cilium project by maintaining best practices for workflow usage, for instance preventing malicious use of GitHub actions. - `@cilium/helm `\_\_: Provide input on the way that Helm can be used to configure features. These owners are a stand-in for the user community to bring a user perspective to the review process. Ensure that Helm changes are defined in manners that will be forward-compatible for upgrade and follow best practices for deployment (for example, being GitOps-friendly). - `@cilium/sig-hubble-api `\_\_: Review Hubble API changes related to gRPC endpoints. The team ensures that API changes are backward compatible or that a new API version is created for backward incompatible changes. - `@cilium/metrics `\_\_: Provide recommendations about the types, names and labels for metrics to follow best practices. This includes considering the cardinality impact of metrics being added or extended. - `@cilium/release-managers `\_\_: Review files related to releases like AUTHORS and VERSION. - `@cilium/security `\_\_: Provide feedback on changes that could have security implications for Cilium, and maintain security-related documentation. - `@cilium/vendor `\_\_: Review vendor updates for software | https://github.com/cilium/cilium/blob/main//Documentation/codeowners.rst | main | cilium | [
-0.07626544684171677,
-0.02616998367011547,
-0.06587637960910797,
-0.016075817868113518,
-0.003192468313500285,
-0.05631878226995468,
-0.03150945156812668,
0.025644317269325256,
0.005788402631878853,
0.05634107440710068,
-0.008107581175863743,
-0.047277048230171204,
-0.03264674171805382,
-... | 0.169876 |
This includes considering the cardinality impact of metrics being added or extended. - `@cilium/release-managers `\_\_: Review files related to releases like AUTHORS and VERSION. - `@cilium/security `\_\_: Provide feedback on changes that could have security implications for Cilium, and maintain security-related documentation. - `@cilium/vendor `\_\_: Review vendor updates for software dependencies to check for any potential upstream breakages / incompatibilities. Discourage the use of unofficial forks of upstream libraries if they are actively maintained. Repository Owners +++++++++++++++++ The following code owners are responsible for a range of general feedback for contributions to specific repositories: - `@cilium/sig-hubble `\_\_: Review all Cilium and Hubble code related to observing system events, exporting those via gRPC protocols outside the node and outside the cluster. those event channels, for example via TLS. - `@cilium/hubble-metrics `\_\_: Review code related to Hubble metrics, ensure changes in exposed metrics are consistent and not breaking without careful consideration. - `@cilium/hubble-ui `\_\_: Maintain the Hubble UI graphical interface. - `@cilium/tetragon `\_\_: Review of all Tetragon code, both for Go and C (for eBPF). The teams above are responsible for reviewing the majority of contributions to the corresponding repositories. Additionally, there are "maintainer" teams listed below which may not be responsible for overall code review for a repository, but they have administrator access to the repositories and so they can assist with configuring GitHub repository settings, secrets, and related processes. For the full codeowners for individual repositories, see the CODEOWNERS file in the corresponding repository. - `@cilium/cilium-cli-maintainers `\_\_ - `@cilium/cilium-maintainers `\_\_ - `@cilium/cilium-packer-ci-build-maintainers `\_\_ - `@cilium/ebpf-lib-maintainers `\_\_ - `@cilium/hubble-maintainers `\_\_ - `@cilium/image-tools-maintainers `\_\_ - `@cilium/metallb-maintainers `\_\_ - `@cilium/openshift-terraform-maintainers `\_\_ - `@cilium/proxy-maintainers `\_\_ - `@cilium/tetragon-maintainers `\_\_ Cloud Integrations ++++++++++++++++++ The following codeowner groups provide insight into the integrations with specific cloud providers: - `@cilium/alibabacloud `\_\_ - `@cilium/aws `\_\_ - `@cilium/azure `\_\_ - `@cilium/gke `\_\_ Cilium Internals ++++++++++++++++ The following codeowner groups cover more specific knowledge about Cilium Agent internals or the way that particular Cilium features interact with external software and protocols: - `@cilium/docker `\_\_: Maintain the deprecated docker-plugin. - `@cilium/endpoint `\_\_: Provide background on how the Cilium Endpoint package fits into the overall agent architecture, relationship with generation of policy / datapath constructs, serialization and restore from disk. - `@cilium/envoy `\_\_: Maintain the L7 proxy integration with Envoy. This includes the configurations for Envoy via xDS protocols as well as the extensible proxylib framework for Go-based layer 7 filters. - `@cilium/egress-gateway `\_\_: Maintain the egress gateway control plane and datapath logic. - `@cilium/fqdn `\_\_: Maintain the L7 DNS proxy integration. - `@cilium/ipcache `\_\_: Provide background on how the userspace IPCache structure fits into the overall agent architecture, ordering constraints with respect to network policies and encryption. Handle the relationship between Kubernetes state and datapath state as it pertains to remote peers. - `@cilium/ipsec `\_\_: Maintain the kernel IPsec configuration and related eBPF logic to ensure traffic is correctly encrypted. - `@cilium/kvstore `\_\_: Review Cilium interactions with key-value stores, particularly etcd. Understand the client libraries used by Cilium for sharing state between nodes and clusters. - `@cilium/loader `\_\_: Maintain the tooling that allows eBPF programs to be loaded into the kernel: LLVM, bpftool, use of cilium/ebpf for loading programs in the agent, ELF templating, etc. - `@cilium/operator `\_\_: Review operations that occur once per cluster via the Cilium Operator component. Take care of the corresponding garbage collection and leader election logic. - `@cilium/proxy `\_\_: Review low-level implementations used to redirect L7 traffic to the actual proxy implementations (FQDN, Envoy, ...). - `@cilium/sig-agent `\_\_: Provide Cilium (agent) general Go review. Internal architecture, core data structures and daemon startup. - `@cilium/sig-bgp `\_\_: Review changes | https://github.com/cilium/cilium/blob/main//Documentation/codeowners.rst | main | cilium | [
-0.019294066354632378,
-0.021248657256364822,
-0.03986276686191559,
-0.018731098622083664,
0.046112678945064545,
-0.05889933928847313,
-0.0552523136138916,
0.045484401285648346,
0.053892143070697784,
-0.00859798863530159,
0.007757280953228474,
-0.06923624128103256,
0.012831803411245346,
-0... | 0.154522 |
of the corresponding garbage collection and leader election logic. - `@cilium/proxy `\_\_: Review low-level implementations used to redirect L7 traffic to the actual proxy implementations (FQDN, Envoy, ...). - `@cilium/sig-agent `\_\_: Provide Cilium (agent) general Go review. Internal architecture, core data structures and daemon startup. - `@cilium/sig-bgp `\_\_: Review changes to our BGP integration. - `@cilium/sig-clustermesh `\_\_: Ensure the reliability of state sharing between clusters to ensure that each cluster maintains a separate fault domain. - `@cilium/sig-datapath `\_\_: Provide feedback on all eBPF code changes, use of the kernel APIs for configuring the networking and socket layers. Coordination of kernel subsystems such as xfrm (IPsec), iptables / nftables, tc. Maintain the control plane layers that populate most eBPF maps; account for endianness and system architecture impacts on the datapath code. - `@cilium/sig-encryption `\_\_ Review control and data plane logic related with encryption (IPSec and WireGuard). - `@cilium/sig-hubble `\_\_: Review all Cilium and Hubble code related to observing system events, exporting those via gRPC protocols outside the node and outside the cluster. Ensure the security of those event channels, for example via TLS. - `@cilium/sig-ipam `\_\_: Coordinate the implementation between all of the IP Address Management modes, provide awareness/insight into IP resource exhaustion and garbage collection concerns. - `@cilium/sig-k8s `\_\_: Provide input on all interactions with Kubernetes, both for standard resources and CRDs. Ensure best practices are followed for the coordination of clusterwide state in order to minimize memory usage. - `@cilium/sig-lb `\_\_: Maintain the layers necessary to coordinate all load balancing configurations within the agent control plane, including Services, ClusterIP, NodePorts, Maglev, local redirect policies, and NAT46/NAT64. - `@cilium/sig-policy `\_\_: Ensure consistency of semantics for all network policy representations. Responsible for all policy logic from Kubernetes down to eBPF policymap entries, including all intermediate layers such as the Policy Repository, SelectorCache, PolicyCache, CachedSelectorPolicy, EndpointPolicy, etc. - `@cilium/sig-scalability `\_\_: Maintain scalability and performance tests. Provide input on scalability and performance related changes. - `@cilium/sig-servicemesh `\_\_: Provide input on the way that Service Mesh constructs such as Gateway API are converted into lower-level constructs backed by eBPF or Envoy configurations. Maintain the CRDs necessary for Service Mesh functionality. - `@cilium/wireguard `\_\_: Maintain the kernel WireGuard configuration and datapath impacts related to ensuring traffic is encrypted correctly when WireGuard mode is enabled. - `@cilium/ztunnel `\_\_: Maintain the ztunnel mTLS-proxy integration in Cilium to ensure the ztunnel agent receives the necessary information to successfully proxy pod-to-pod traffic through mTLS tunnels. | https://github.com/cilium/cilium/blob/main//Documentation/codeowners.rst | main | cilium | [
-0.0020590866915881634,
-0.041078079491853714,
-0.04222873970866203,
-0.003565068356692791,
0.001266168081201613,
0.013701207004487514,
-0.04092542082071304,
0.05316701531410217,
-0.07944586873054504,
-0.0037907669320702553,
-0.03415302187204361,
-0.05550343543291092,
-0.0026747677475214005,... | 0.222515 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_api\_ref: ############# API Reference ############# \*\*\*\*\*\*\*\*\*\*\*\* Introduction \*\*\*\*\*\*\*\*\*\*\*\* The Cilium API is JSON based and provided by the ``cilium-agent``. The purpose of the API is to provide visibility and control over an individual agent instance. In general, all API calls affect only the resources managed by the individual ``cilium-agent`` serving the API. A few selected API calls such as the security identity resolution provides cluster wide visibility. Such API calls are marked specifically. Unless noted otherwise, API calls will only affect local agent resources. \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* How to access the API \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* CLI Client ========== The easiest way to access the API is via the ``cilium`` CLI client. ``cilium`` will automatically locate the API of the agent running on the same node and access it. However, using the ``-H`` or ``--host`` flag, the ``cilium`` client can be pointed to an arbitrary API address. Example ------- .. code-block:: shell-session $ cilium-dbg -H unix:///var/run/cilium/cilium.sock [...] Golang Package ============== The following Go packages can be used to access the API: +---------------------+---------------------------------------------------------------+ | Package | Description | +---------------------+---------------------------------------------------------------+ | `pkg/client`\_ | Main client API abstraction | +---------------------+---------------------------------------------------------------+ | `api/v1/models`\_ | API resource data type models | +---------------------+---------------------------------------------------------------+ Example ------- The full example can be found in the `cilium/client-example`\_ repository. .. code-block:: go import ( "fmt" "github.com/cilium/cilium/pkg/client" ) func main() { c, err := client.NewDefaultClient() if err != nil { ... } endpoints, err := c.EndpointList() if err != nil { ... } for \_, ep := range endpoints { fmt.Printf("%8d %14s %16s %32s\n", ep.ID, ep.ContainerName, ep.Addressing.IPV4, ep.Addressing.IPV6) } \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Compatibility Guarantees \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Cilium API is stable as of version 1.0, backward compatibility will be upheld for whole lifecycle of Cilium 1.x. \*\*\*\*\*\*\*\*\*\*\*\*\* API Reference \*\*\*\*\*\*\*\*\*\*\*\*\* .. openapi:: ../api/v1/openapi.yaml .. \_pkg/client: https://godoc.org/github.com/cilium/cilium/pkg/client .. \_api/v1/models: https://godoc.org/github.com/cilium/cilium/api/v1/models .. \_cilium/client-example: https://github.com/cilium/client-example | https://github.com/cilium/cilium/blob/main//Documentation/api.rst | main | cilium | [
-0.04658131301403046,
0.05209015682339668,
-0.097252756357193,
-0.007760847453027964,
0.059780098497867584,
-0.06476884335279465,
-0.028319096192717552,
0.011140009388327599,
0.06248059868812561,
-0.04300794005393982,
0.0371241457760334,
-0.0666016936302185,
0.04331293702125549,
-0.0305904... | 0.212409 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_glossary: Glossary ======== Cilium has some terms with special meanings. These should all be covered throughout the documentation but for convenience we have also listed some of them below with short descriptions. If you need more information, please ask us on `Cilium Slack`\_. Feel free to extend this document with words you expected to see here. .. glossary:: :sorted: Endpoint :ref:`endpoint` Policy A Cilium policy consists of a list of rules. The security policy can be specified in The Kubernetes NetworkPolicy format or The Cilium policy language. iproute2 https://www.kernel.org/pub/linux/utils/net/iproute2/ llvm https://releases.llvm.org/ Linux kernel https://www.kernel.org/ DaemonSet https://kubernetes.io/docs/concepts/workloads/controllers/daemonset/ ConfigMap https://kubernetes.io/docs/tasks/configure-pod-container/configure-pod-configmap/ CNI https://github.com/containernetworking/cni RBAC https://kubernetes.io/docs/reference/access-authn-authz/rbac/ NodeSelector https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector Volumes https://kubernetes.io/docs/tasks/configure-pod-container/configure-volume-storage/ VXLAN https://tools.ietf.org/html/rfc7348 Geneve https://tools.ietf.org/html/rfc8926 Pod https://kubernetes.io/docs/concepts/workloads/pods/pod/ Service https://kubernetes.io/docs/concepts/services-networking/service/ CustomResourceDefinition https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/#customresourcedefinitions HeadlessServices https://kubernetes.io/docs/concepts/services-networking/service/#headless-services Helm https://helm.sh/ | https://github.com/cilium/cilium/blob/main//Documentation/glossary.rst | main | cilium | [
0.018881579861044884,
0.01707490347325802,
-0.03881723806262016,
-0.061911340802907944,
-0.014231482520699501,
-0.02997736446559429,
-0.012735256925225258,
0.042067818343639374,
0.035141199827194214,
0.015589621849358082,
0.035880185663700104,
-0.01300786528736353,
0.013912705704569817,
-0... | 0.230356 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_verify\_image\_signatures: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Verifying Image Signatures \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Prerequisites ============= You will need to `install cosign`\_. .. \_`install cosign`: https://docs.sigstore.dev/cosign/installation/ Verify Signed Container Images ============================== Since version 1.13, all Cilium container images are signed using cosign. Let's verify a Cilium image's signature using the ``cosign verify`` command: .. code-block:: shell-session $ TAG=v1.13.0 $ cosign verify --certificate-github-workflow-repository cilium/cilium \ --certificate-oidc-issuer https://token.actions.githubusercontent.com \ --certificate-github-workflow-name "Image Release Build" \ --certificate-github-workflow-ref refs/tags/${TAG} \ --certificate-identity "https://github.com/cilium/cilium/.github/workflows/build-images-releases.yaml@refs/tags/${TAG}" \ "quay.io/cilium/cilium:${TAG}" | jq .. note:: ``cosign`` is used to verify images signed in ``KEYLESS`` mode. To learn more about keyless signing, please refer to `Keyless Signatures`\_. ``--certificate-github-workflow-name string`` contains the workflow claim from the GitHub OIDC Identity token that contains the name of the executed workflow. For the names of workflows used to build Cilium images, see the ``build-images`` workflows under `Cilium workflows`\_. ``--certificate-github-workflow-ref string`` contains the ref claim from the GitHub OIDC Identity token that contains the git ref that the workflow run was based upon. ``--certificate-identity`` is used to verify the identity of the certificate from the GitHub build images release workflow. .. \_`Keyless Signatures`: https://docs.sigstore.dev/cosign/signing/overview/#the-signing-witnessing-and-verifying-process .. \_`Cilium workflows`: https://github.com/cilium/cilium/tree/main/.github/workflows | https://github.com/cilium/cilium/blob/main//Documentation/configuration/verify-image-signatures.rst | main | cilium | [
-0.01686711050570011,
0.06927002221345901,
-0.0524456612765789,
-0.03183647245168686,
0.08059440553188324,
-0.08738638460636139,
-0.02058529481291771,
0.0359770692884922,
0.021620802581310272,
-0.022897467017173767,
0.09075397253036499,
-0.0732874795794487,
0.03480018302798271,
-0.02775341... | 0.123829 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_configuration: Configuration ============= Your Cilium installation is configured by one or more Helm values - see :ref:`helm\_reference`. These helm values are converted to arguments for the individual components of a Cilium installation, such as :doc:`../cmdref/cilium-agent` and :doc:`../cmdref/cilium-operator`, and stored in a ConfigMap. .. \_cilium-config-configmap: ``cilium-config`` ConfigMap ----------------------------- These arguments are stored in a shared ConfigMap called ``cilium-config`` (albeit without the leading ``--``). For example, a typical installation may look like .. code-block:: shell-session $ kubectl -n kube-system get configmap cilium-config -o yaml data: agent-not-ready-taint-key: node.cilium.io/agent-not-ready auto-direct-node-routes: "false" (output continues) .. \_making-config-changes: Making Changes -------------- You may change the configuration of a running installation in three ways: #. Via ``helm upgrade`` Do so by providing new values to Helm and applying them to the existing installation. By setting the value ``rollOutCiliumPods=true``, the agent pods will be gradually restarted. #. Via ``cilium config set`` The `Cilium CLI `\_ has the ability to update individual values in the ``cilium-config`` ConfigMap. By default Cilium Agent pods are restarted when configuration is changed. To gradually restart do ``cilium config set --restart=false ...`` and manually delete agent pods to pick up the changes. #. Via ``CiliumNodeConfig`` objects Cilium also supports configuration on sets of nodes. See the :ref:`per-node-configuration` page for more details. This requires that pods be manually deleted for changes to take effect. Core Agent ---------- .. toctree:: :maxdepth: 1 :glob: api-rate-limiting api-restrictions per-node-config sctp vlan-802.1q argocd-issues Security -------- .. toctree:: :maxdepth: 1 :glob: verify-image-signatures sbom | https://github.com/cilium/cilium/blob/main//Documentation/configuration/index.rst | main | cilium | [
0.015643002465367317,
0.029241008684039116,
-0.09153801947832108,
-0.015021809376776218,
0.038198839873075485,
-0.05678120627999306,
-0.04047852382063866,
0.056294891983270645,
0.03466878458857536,
-0.06292484700679779,
0.08028753101825714,
-0.08986285328865051,
0.05541222542524338,
-0.063... | 0.148224 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_sctp: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* SCTP support (beta) \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* .. include:: ../beta.rst Enabling ======== Pass ``--set sctp.enabled=true`` to helm. .. admonition:: Video :class: attention You can also watch a video explanation of Cilium's SCTP support in `eCHO episode 78: Stream Control Transmission Protocol (SCTP) `\_\_. Limitations =========== Cilium supports basic SCTP support. Specifically, the following is supported: - Pod <-> Pod communication - Pod <-> Service communication [\*] - Pod <-> Pod communication with network policies applied to SCTP traffic [\*] .. note:: [\*] SCTP support does not support rewriting ports for SCTP packets. This means that when defining services, the targetPort \*\*MUST\*\* equal the port, otherwise the packet will be dropped. .. warning:: Cilium does not support the following for SCTP: - Multihoming - Policies for pod-to-VIP - :ref:`Kube-proxy replacement (KPR)` when port rewriting is necessary: for example, NodePort Services are not supported with the combination of KPR and SCTP. - BPF masquerading - Egress gateway | https://github.com/cilium/cilium/blob/main//Documentation/configuration/sctp.rst | main | cilium | [
0.010904966853559017,
0.016131550073623657,
-0.07748745381832123,
-0.03425055742263794,
0.09281326085329056,
0.025346292182803154,
0.0243584793061018,
0.010791933164000511,
0.023567289113998413,
-0.02423354610800743,
0.055073536932468414,
-0.04461295157670975,
0.00330072408542037,
-0.00956... | 0.219412 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Administrative API Enablement \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Cilium 1.14 introduced a new set of flags that you can use to selectively enable which API endpoints are exposed to clients. When an API client makes a request to an API endpoint that is administratively disabled, the server responds with an HTTP 403 Forbidden error. You can configure the option with a list of endpoints as described in the following sections, or by specifying an option with the ``\*`` suffix. If ``\*`` is provided directly as a flag value, then all APIs are enabled. If there is text before the ``\*``, then the API flag must start with that prefix in order for the flag to enable that option. For example, ``Get\*`` enables all read-only "GET" APIs without enabling any write APIs. The cilium-agent relies on several of these APIs for its basic duties. In particular, disabling the following APIs will likely cause significant disruption to agent operations: - ``GetConfig`` - ``GetHealthz`` - ``PutEndpointID`` - ``DeleteEndpointID`` - ``PostIPAM`` - ``DeleteIPAMIP`` The following sections outline the flags for different Cilium binaries and the API endpoints that may be configured using those flags. .. include:: api-restrictions-table.rst | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-restrictions.rst | main | cilium | [
-0.026615507900714874,
0.07102186232805252,
-0.09570915997028351,
-0.038783736526966095,
0.052013099193573,
-0.06244715303182602,
-0.06632272899150848,
0.005641862750053406,
0.037274111062288284,
-0.0151714151725173,
0.02603248506784439,
-0.03841989114880562,
0.023974010720849037,
0.021468... | 0.182272 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_vlan\_802.1q: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* VLAN 802.1q support \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Cilium enables firewalling on native devices in use and will filter all unknown traffic. VLAN 802.1q packets will always be passed through their main device with associated tag (e.g. VLAN device is ``eth0.4000`` and its main interface is ``eth0``). By default, Cilium will allow all tags from the native devices (i.e. if ``eth0.4000`` is controlled by Cilium and has an eBPF program attached, then VLAN tag ``4000`` will be allowed on device ``eth0``). Additional VLAN tags may be allowed with the cilium-agent flag ``--vlan-bpf-bypass=4001,4002`` (or Helm variable ``--set bpf.vlanBypass="[4001,4002]"``). The list of allowed VLAN tags cannot be too big in order to keep eBPF program of predictable size. Currently this list should contain no more than 5 entries. If you need more, then there is only one way for now: you need to allow all tags with cilium-agent flag ``--vlan-bpf-bypass=0``. .. note:: Currently, the cilium-agent will scan for available VLAN devices and tags only on startup. | https://github.com/cilium/cilium/blob/main//Documentation/configuration/vlan-802.1q.rst | main | cilium | [
0.010260284878313541,
0.07263986766338348,
-0.11901986598968506,
-0.0519547276198864,
0.07262452691793442,
-0.08592385053634644,
0.011354513466358185,
-0.0022773281671106815,
-0.03795574977993965,
-0.05077202618122101,
0.07609152048826218,
-0.0844007059931755,
0.05150933936238289,
-0.01138... | 0.170746 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_sbom: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Software Bill of Materials \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* A Software Bill of Materials (SBOM) is a complete, formally structured list of components that are required to build a given piece of software. SBOM provides insight into the software supply chain and any potential concerns related to license compliance and security that might exist. The Cilium SBOM is generated using the `syft`\_ tool. To learn more about SBOM, see `what is an SBOM`\_. .. \_`syft`: https://github.com/anchore/syft .. \_`what is an SBOM`: https://edu.chainguard.dev/open-source/sbom/what-is-an-sbom/ Prerequisites ============= - `Install cosign`\_ .. \_`Install cosign`: https://docs.sigstore.dev/cosign/system\_config/installation/ Download SBOM ============= You can download the SBOM in-toto attestation from the supplied Cilium image using the following command: .. code-block:: shell-session $ cosign download attestation --predicate-type spdxjson | jq -r .payload | base64 -d | jq .predicate > ciliumSBOM.spdx.json Verify SBOM attestation ======================= To verify the SBOM in-toto attestation on the supplied Cilium image, run the following command: .. parsed-literal:: $ TAG=|IMAGE\_TAG| $ cosign verify-attestation --certificate-github-workflow-repository cilium/cilium \ --certificate-oidc-issuer https://token.actions.githubusercontent.com \ --certificate-identity-regexp https://github.com/cilium/cilium/.github/workflows \ --type spdxjson 2>&1 | head -n 13 For example: .. code-block:: shell-session $ cosign verify-attestation --certificate-github-workflow-repository cilium/cilium \ --certificate-oidc-issuer https://token.actions.githubusercontent.com \ --certificate-identity-regexp https://github.com/cilium/cilium/.github/workflows \ --type spdxjson quay.io/cilium/cilium-ci:d2d270a42b674ca1e7c536186691d8ac8317fd64 2>&1 | head -n 13 Verification for quay.io/cilium/cilium-ci:d2d270a42b674ca1e7c536186691d8ac8317fd64 -- The following checks were performed on each of these signatures: - The cosign claims were validated - Existence of the claims in the transparency log was verified offline - The code-signing certificate was verified using trusted certificate authority certificates Certificate subject: https://github.com/cilium/cilium/.github/workflows/build-images-ci.yaml@refs/pull/34011/merge Certificate issuer URL: https://token.actions.githubusercontent.com GitHub Workflow Trigger: pull\_request GitHub Workflow SHA: 7d967b8355489cef6a787558ac70c9c619463284 GitHub Workflow Name: Image CI Build GitHub Workflow Repository: cilium/cilium GitHub Workflow Ref: refs/pull/34011/merge It can be validated that the image was signed using GitHub Actions in the Cilium repository from the ``Certificate subject`` and ``Certificate issuer URL`` fields of the output. .. note:: The `in-toto`\_ Attestation Framework provides a specification for generating verifiable claims about any aspect of how a piece of software is produced. Consumers or users of software can then validate the origins of the software, and establish trust in its supply chain, using in-toto attestations. .. \_`in-toto`: https://in-toto.io/ | https://github.com/cilium/cilium/blob/main//Documentation/configuration/sbom.rst | main | cilium | [
-0.05769704282283783,
-0.03062795102596283,
-0.11694851517677307,
-0.0641418844461441,
0.044647589325904846,
-0.04276961460709572,
0.04183375462889671,
0.052862074226140976,
0.008340010419487953,
-0.005034903530031443,
0.05612611398100853,
0.008592645637691021,
0.04585357382893562,
-0.0430... | 0.297716 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_api\_rate\_limiting: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* API Rate Limiting \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* The per node Cilium agent is essentially event-driven. For example, the CNI plugin is invoked when a new workload is scheduled onto the node which in turn makes an API call to the Cilium agent to allocate an IP address and create the Cilium endpoint. Another example is loading of network policy or service definitions where changes of these definitions will create an event which will notify the Cilium agent that a modification is required. Due to being event-driven, the amount of work performed by the Cilium agent highly depends on the rate of external events it receives. In order to constrain the resources that the Cilium agent consumes, it can be helpful to restrict the rate and allowed parallel executions of API calls. Default Rate Limits =================== The following API calls are currently subject to rate limiting: ========================== ====== ===== ============= ============ ================= =========== =============================== API Call Limit Burst Max Parallel Min Parallel Max Wait Duration Auto Adjust Estimated Processing Duration ========================== ====== ===== ============= ============ ================= =========== =============================== ``PUT /endpoint/{id}`` 0.5/s 4 4 15s True 2s ``DELETE /endpoint/{id}`` 4 4 True 200ms ``GET /endpoint/{id}/\*`` 4/s 4 4 2 10s True 200ms ``PATCH /endpoint/{id}\*`` 0.5/s 4 4 15s True 1s ``GET /endpoint`` 1/s 4 2 2 True 300ms ========================== ====== ===== ============= ============ ================= =========== =============================== Configuration ============= The ``api-rate-limit`` option can be used to overwrite individual settings of the default configuration:: --api-rate-limit endpoint-create=rate-limit:2/s,rate-burst:4 API call to Configuration mapping --------------------------------- ========================== ==================== API Call Config Name ========================== ==================== ``PUT /endpoint/{id}`` ``endpoint-create`` ``DELETE /endpoint/{id}`` ``endpoint-delete`` ``GET /endpoint/{id}/\*`` ``endpoint-get`` ``PATCH /endpoint/{id}\*`` ``endpoint-patch`` ``GET /endpoint`` ``endpoint-list`` ========================== ==================== Configuration Parameters ------------------------ ================================= ========= ========= ===================================================================================== Configuration Key Example Default Description ================================= ========= ========= ===================================================================================== ``rate-limit`` ``5/m`` None Allowed requests per time unit in the format ``/``. ``rate-burst`` ``4`` None Burst of API requests allowed by rate limiter. ``min-wait-duration`` ``10ms`` ``0`` Minimum wait duration each API call has to wait before being processed. ``max-wait-duration`` ``15s`` ``0`` Maximum duration an API call is allowed to wait before it fails. ``estimated-processing-duration`` ``100ms`` ``0`` Estimated processing duration of an average API call. Used for automatic adjustment. ``auto-adjust`` ``true`` ``false`` Enable automatic adjustment of ``rate-limit``, ``rate-burst`` and ``parallel-requests``. ``parallel-requests`` ``4`` ``0`` Number of parallel API calls allowed. ``min-parallel-requests`` ``2`` ``0`` Lower limit of parallel requests when auto-adjusting. ``max-parallel-requests`` ``6`` ``0`` Upper limit of parallel requests when auto-adjusting. ``mean-over`` ``10`` ``10`` Number of API calls to calculate mean processing duration for auto adjustment. ``log`` ``true`` ``false`` Log an Info message for each API call processed. ``delayed-adjustment-factor`` ``0.25`` ``0.5`` Factor for slower adjustment of ``rate-burst`` and ``parallel-requests``. ``max-adjustment-factor`` ``10.0`` ``100.0`` Maximum factor the auto-adjusted values can deviate from the initial base values configured. ================================= ========= ========= ===================================================================================== Valid duration values --------------------- The ``rate-limit`` option expects a value in the form ``/`` where ```` is a value that can be parsed with `ParseDuration() `\_. The supported units are: ``ns``, ``us``, ``ms``, ``s``, ``m``, ``h``. \*\*Examples:\*\* \* ``rate-limit:10/2m`` \* ``rate-limit:3.5/h`` \* ``rate-limit:1/100ms`` Automatic Adjustment ==================== Static values are relatively useless as the Cilium agent will run on different machine types. Deriving rate limits based on number of available CPU cores or available memory can be misleading as well as the Cilium agent may be subject to CPU and memory constraints. For this reason, all API call rate limiting is done with automatic adjustment of the limits with the goal to stay as close as possible to the configured estimated | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-rate-limiting.rst | main | cilium | [
-0.05270133912563324,
0.01800060085952282,
-0.06663168966770172,
-0.009005239233374596,
0.02269173040986061,
-0.05025395750999451,
-0.03199247643351555,
0.01823108270764351,
0.10234876722097397,
0.0025902458000928164,
0.02939928136765957,
-0.06114308908581734,
0.03822949528694153,
-0.03634... | 0.248343 |
cores or available memory can be misleading as well as the Cilium agent may be subject to CPU and memory constraints. For this reason, all API call rate limiting is done with automatic adjustment of the limits with the goal to stay as close as possible to the configured estimated processing duration. This processing duration is specified for each group of API call and is constantly monitored. On completion of every API call, new limits are calculated. For this purpose, an adjustment factor is calculated: .. code-block:: go AdjustmentFactor := EstimatedProcessingDuration / MeanProcessingDuration AdjustmentFactor = Min(Max(AdjustmentFactor, 1.0/MaxAdjustmentFactor), MaxAdjustmentFactor) This adjustment factor is then applied to ``rate-limit``, ``rate-burst`` and ``parallel-requests`` and will steer the mean processing duration to get closer to the estimated processing duration. If ``delayed-adjustment-factor`` is specified, then this additional factor is used to slow the growth of the ``rate-burst`` and ``parallel-requests`` as both values should typically adjust slower than ``rate-limit``: .. code-block:: go NewValue = OldValue \* AdjustmentFactor NewValue = OldValue + ((NewValue - OldValue) \* DelayedAdjustmentFactor) Metrics ======= All API calls subject to rate limiting will expose :ref:`metrics\_api\_rate\_limiting`. Example:: cilium\_api\_limiter\_adjustment\_factor api\_call="endpoint-create" 0.695787 cilium\_api\_limiter\_processed\_requests\_total api\_call="endpoint-create" outcome="success" return\_code="200" 7.000000 cilium\_api\_limiter\_processing\_duration\_seconds api\_call="endpoint-create" value="estimated" 2.000000 cilium\_api\_limiter\_processing\_duration\_seconds api\_call="endpoint-create" value="mean" 2.874443 cilium\_api\_limiter\_rate\_limit api\_call="endpoint-create" value="burst" 4.000000 cilium\_api\_limiter\_rate\_limit api\_call="endpoint-create" value="limit" 0.347894 cilium\_api\_limiter\_requests\_in\_flight api\_call="endpoint-create" value="in-flight" 0.000000 cilium\_api\_limiter\_requests\_in\_flight api\_call="endpoint-create" value="limit" 0.000000 cilium\_api\_limiter\_wait\_duration\_seconds api\_call="endpoint-create" value="max" 15.000000 cilium\_api\_limiter\_wait\_duration\_seconds api\_call="endpoint-create" value="mean" 0.000000 cilium\_api\_limiter\_wait\_duration\_seconds api\_call="endpoint-create" value="min" 0.000000 Understanding the log output ============================ The API rate limiter logs under the ``rate`` subsystem. An example message can be seen below:: level=info msg="API call has been processed" name=endpoint-create processingDuration=772.847247ms subsys=rate totalDuration=14.923958916s uuid=d34a2e1f-1ac9-11eb-8663-42010a8a0fe1 waitDurationTotal=14.151023084s The following is an explanation for all the API rate limiting messages: :: "Processing API request with rate limiter" The request was admitted into the rate limiter. The associated HTTP context (caller's request) has not yet timed out. The request will now be rate-limited according to the configuration of the rate limiter. It will enter the waiting stage according to the computed waiting duration. :: "API request released by rate limiter" The request has finished waiting its computed duration to achieve rate-limiting. The underlying HTTP API action will now take place. This means that this request was not thrown back at the caller with a 429 HTTP status code. This is a common message when the requests are being processed within the configured bounds of the rate limiter. :: "API call has been processed": The API rate limiter has processed this request and the underlying HTTP API action has finished. This means the request is no longer actively waiting or in other words, no longer being rate-limited. This does not mean the underlying HTTP action has succeeded; only that this request has been dealt with. :: "Not processing API request due to cancelled context" The underlying HTTP context (request) was cancelled. In other words, the caller has given up on the request. This most likely means that the HTTP request timed out. A 429 HTTP response status code is returned to the caller, which may or may not receive it anyway. :: "Not processing API request. Wait duration for maximum parallel requests exceeds maximum" The request has been denied by the rate limiter because too many parallel requests are already in flight. The caller will receive a 429 HTTP status response. This is a common message when the rate limiter is doing its job of preventing too many parallel requests at once. :: "Not processing API request. Wait duration exceeds maximum" The request has been denied by the rate limiter because the request's waiting duration would exceed the maximum configured waiting duration. For example, if the maximum waiting duration was ``5s`` | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-rate-limiting.rst | main | cilium | [
-0.010313470847904682,
0.007400126196444035,
-0.07417433708906174,
0.004701122175902128,
-0.04531801864504814,
-0.08875949680805206,
-0.12142475694417953,
0.06501100212335587,
0.03318620100617409,
-0.01368282362818718,
0.05940893664956093,
-0.03197868540883064,
0.008573593571782112,
-0.071... | 0.163462 |
doing its job of preventing too many parallel requests at once. :: "Not processing API request. Wait duration exceeds maximum" The request has been denied by the rate limiter because the request's waiting duration would exceed the maximum configured waiting duration. For example, if the maximum waiting duration was ``5s`` and due to the backlog of the rate limiter, this request would need to wait ``10s``, then this request would be thrown out. A 429 HTTP response status code would be returned to the caller. This is the most common message when the rate limiter is doing its job of pacing the incoming requests into Cilium. :: "Not processing API request due to cancelled context while waiting" The request has been denied by the rate limiter because after the request has waited its calculated waiting duration, the context associated with the request has been cancelled. In the most likely scenario, this means that there was an HTTP timeout while the request was actively being rate-limited or in other words, actively being delayed by the rate limiter. A 429 HTTP response status code is returned to the caller. | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-rate-limiting.rst | main | cilium | [
-0.07756512612104416,
0.029780443757772446,
-0.07251645624637604,
-0.021251194179058075,
-0.02306702360510826,
-0.13677944242954254,
-0.08925648033618927,
-0.007471657823771238,
0.08555377274751663,
0.012229940854012966,
0.06661710888147354,
-0.04587029293179512,
-0.012779354117810726,
-0.... | 0.166873 |
.. Cilium Agent API ================ The following API flags are compatible with the ``cilium-agent`` flag ``enable-cilium-api-server-access``. ===================== ==================== Flag Name Description ===================== ==================== DeleteEndpoint Deletes a list of endpoints that have endpoints matching the provided properties DeleteEndpointID Deletes the endpoint specified by the ID. Deletion is imminent and atomic, if the deletion request is valid and the endpoint exists, deletion will occur even if errors are encountered in the process. If errors have been encountered, the code 202 will be returned, otherwise 200 on success. All resources associated with the endpoint will be freed and the workload represented by the endpoint will be disconnected.It will no longer be able to initiate or receive communications of any sort. DeleteFqdnCache Deletes matching DNS lookups from the cache, optionally restricted by DNS name. The removed IP data will no longer be used in generated policies. DeleteIPAMIP - DeletePrefilter - GetBGPPeers Retrieves current operational state of BGP peers created by Cilium BGP virtual router. This includes session state, uptime, information per address family, etc. GetBGPRoutePolicies Retrieves route policies from BGP Control Plane. GetBGPRoutes Retrieves routes from BGP Control Plane RIB filtered by parameters you specify GetCgroupDumpMetadata - GetClusterNodes - GetConfig Returns the configuration of the Cilium daemon. GetDebuginfo - GetEndpoint Retrieves a list of endpoints that have metadata matching the provided parameters, or all endpoints if no parameters provided. GetEndpointID Returns endpoint information GetEndpointIDConfig Retrieves the configuration of the specified endpoint. GetEndpointIDHealthz - GetEndpointIDLabels - GetEndpointIDLog - GetFqdnCache Retrieves the list of DNS lookups intercepted from endpoints, optionally filtered by DNS name, CIDR IP range or source. GetFqdnCacheID Retrieves the list of DNS lookups intercepted from the specific endpoint, optionally filtered by endpoint id, DNS name, CIDR IP range or source. GetFqdnNames Retrieves the list of DNS-related fields (names to poll, selectors and their corresponding regexes). GetHealthz Returns health and status information of the Cilium daemon and related components such as the local container runtime, connected datastore, Kubernetes integration and Hubble. GetIP Retrieves a list of IPs with known associated information such as their identities, host addresses, Kubernetes pod names, etc. The list can optionally filtered by a CIDR IP range. GetIdentity Retrieves a list of identities that have metadata matching the provided parameters, or all identities if no parameters are provided. GetIdentityEndpoints - GetIdentityID - GetLRP - GetMap - GetMapName - GetMapNameEvents - GetNodeIds Retrieves a list of node IDs allocated by the agent and their associated node IP addresses. GetPolicy Returns the entire policy tree with all children. Deprecated: will be removed in v1.19 GetPolicySelectors - GetPrefilter - GetService - PatchConfig Updates the daemon configuration by applying the provided ConfigurationMap and regenerates & recompiles all required datapath components. PatchEndpointID Applies the endpoint change request to an existing endpoint PatchEndpointIDConfig Update the configuration of an existing endpoint and regenerates & recompiles the corresponding programs automatically. PatchEndpointIDLabels Sets labels associated with an endpoint. These can be user provided or derived from the orchestration system. PatchPrefilter - PostIPAM - PostIPAMIP - PutEndpointID Creates a new endpoint ===================== ==================== Cilium Agent Clusterwide Health API =================================== The following API flags are compatible with the ``cilium-agent`` flag ``enable-cilium-health-api-server-access``. ===================== ==================== Flag Name Description ===================== ==================== GetHealthz Returns health and status information of the local node including load and uptime, as well as the status of related components including the Cilium daemon. GetStatus Returns the connectivity status to all other cilium-health instances using interval-based probing. PutStatusProbe Runs a synchronous probe to all other cilium-health instances and returns the connectivity status. ===================== ==================== Cilium Operator API =================== The following API flags are compatible with the ``cilium-operator`` flag ``enable-cilium-operator-server-access``. ===================== ==================== | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-restrictions-table.rst | main | cilium | [
-0.04431693255901337,
0.003929180093109608,
-0.11107897013425827,
0.005523163825273514,
-0.018036970868706703,
-0.06185239925980568,
-0.006801071111112833,
-0.062445465475320816,
0.07561546564102173,
-0.031119773164391518,
0.024239487946033478,
-0.03667304664850235,
0.036695025861263275,
-... | 0.217737 |
the Cilium daemon. GetStatus Returns the connectivity status to all other cilium-health instances using interval-based probing. PutStatusProbe Runs a synchronous probe to all other cilium-health instances and returns the connectivity status. ===================== ==================== Cilium Operator API =================== The following API flags are compatible with the ``cilium-operator`` flag ``enable-cilium-operator-server-access``. ===================== ==================== Flag Name Description ===================== ==================== GetCluster Returns the list of remote clusters and their status. GetHealthz Returns the status of cilium operator instance. GetMetrics Returns the metrics exposed by the Cilium operator. ===================== ==================== | https://github.com/cilium/cilium/blob/main//Documentation/configuration/api-restrictions-table.rst | main | cilium | [
0.014606880955398083,
0.01807682402431965,
-0.11901342123746872,
0.019251558929681778,
-0.019894247874617577,
-0.09746622294187546,
-0.04300383850932121,
-0.024190299212932587,
-0.023507844656705856,
-0.030327534303069115,
0.054948948323726654,
-0.058654677122831345,
0.020960062742233276,
... | 0.160114 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_argocd\_issues: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Troubleshooting Cilium deployed with Argo CD \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* There have been reports from users hitting issues with Argo CD. This documentation page outlines some of the known issues and their solutions. Argo CD deletes Cilium custom resources ======================================= When deploying Cilium with Argo CD, some users have reported that Cilium-generated custom resources disappear, causing the following issues: - ``ciliumid`` not found (:gh-issue:`17614`) Solution -------- To prevent this issue, declare resource exclusions in the Argo CD ``ConfigMap`` by following `these instructions `\_\_. Here is an example snippet: .. code-block:: yaml resource.exclusions: | - apiGroups: - cilium.io kinds: - CiliumIdentity clusters: - "\*" Argo CD show resources permanently out-of-sync ============================================== - Argo CD Out-of-sync issues for hubble-generate-certs (:gh-issue:`14550`) - Out-of-sync issues for Cilium using Argo CD (:gh-issue:`18298`) Solution -------- You may pick one of the following approaches. The ``argocd.argoproj.io/compare-options: IgnoreExtraneous`` annotation can be added to resources, which use non-idempotent helm generators to avoid this issue. The helm value ``nonIdempotentAnnotations`` is available for the purpose and can be set in your values file. .. code-block:: yaml nonIdempotentAnnotations: argocd.argoproj.io/compare-options: IgnoreExtraneous Exclusions can be set in your Argo CD application definition to avoid getting “out of sync” when certificates get regenerated. The example below is for Hubble, similar secrets exist however for clustermesh as well. They are all labeled with ``cilium.io/helm-template-non-idempotent: "true"``. .. code-block:: yaml ignoreDifferences: - group: "" kind: ConfigMap name: hubble-ca-cert jsonPointers: - /data/ca.crt - group: "" kind: Secret name: hubble-relay-client-certs jsonPointers: - /data/ca.crt - /data/tls.crt - /data/tls.key - group: "" kind: Secret name: hubble-server-certs jsonPointers: - /data/ca.crt - /data/tls.crt - /data/tls.key .. note:: After applying the above configurations, for the settings to take effect, you will need to restart the Argo CD deployments. Helm template with serviceMonitor enabled fails =============================================== Some users have reported that when they install Cilium using Argo CD and run ``helm template`` with ``serviceMonitor`` enabled, it fails. It fails because Argo CD CLI doesn't pass the ``--api-versions`` flag to Helm upon deployment. Solution -------- This `pull request `\_\_ fixed this issue in Argo CD's `v2.3.0 release `\_\_. Upgrade your Argo CD and check if ``helm template`` with ``serviceMonitor`` enabled still fails. .. note:: When using ``helm template``, it is highly recommended you set ``--kube-version`` and ``--api-versions`` with the values matching your target Kubernetes cluster. Helm charts such as Cilium's often conditionally enable certain Kubernetes features based on their availability (beta vs stable) on the target cluster. By specifying ``--api-versions=monitoring.coreos.com/v1`` you should be able to pass validation with ``helm template``. If you have an issue with Argo CD that's not outlined above, check this `list of Argo CD related issues on GitHub `\_\_. If you can't find an issue that relates to yours, create one and/or seek help on `Cilium Slack`\_. Application chart for Cilium deployed to Talos Linux fails with: field not declared in schema ============================================================================================= When deploying Cilium to Talos Linux with ArgoCD, some users have reported issues due to Talos Security configuration. ArgoCD may fail to deploy the application with the message:: Failed to compare desired state to live state: failed to calculate diff: error calculating structured merge diff: error building typed value from live resource: .spec.template.spec.securityContext.appArmorProfile: field not declared in schema Solution -------- Add option ``ServerSideApply=true`` to list ``syncPolicy.syncOptions`` for the Application. .. code-block:: yaml apiVersion: argoproj.io/v1alpha1 kind: Application spec: syncPolicy: syncOptions: - ServerSideApply=true Visit the `ArgoCD documentation `\_\_ for further details. | https://github.com/cilium/cilium/blob/main//Documentation/configuration/argocd-issues.rst | main | cilium | [
-0.023538930341601372,
0.03382917493581772,
-0.09694694727659225,
-0.0448879636824131,
0.004916099831461906,
-0.07042091339826584,
-0.039533283561468124,
0.015907950699329376,
-0.01200197171419859,
0.02174803428351879,
0.06351300328969955,
-0.039842747151851654,
-0.0067987702786922455,
-0.... | 0.204077 |
declared in schema Solution -------- Add option ``ServerSideApply=true`` to list ``syncPolicy.syncOptions`` for the Application. .. code-block:: yaml apiVersion: argoproj.io/v1alpha1 kind: Application spec: syncPolicy: syncOptions: - ServerSideApply=true Visit the `ArgoCD documentation `\_\_ for further details. | https://github.com/cilium/cilium/blob/main//Documentation/configuration/argocd-issues.rst | main | cilium | [
-0.041181620210409164,
-0.06233767047524452,
-0.12037158757448196,
-0.015932364389300346,
-0.017410818487405777,
-0.03490624949336052,
-0.00027017146931029856,
-0.05903371050953865,
-0.033893924206495285,
0.0767960250377655,
0.05195809528231621,
0.014518975280225277,
-0.04450423642992973,
... | 0.110302 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_per-node-configuration: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Per-node configuration \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* The Cilium agent process (a.k.a. DaemonSet) supports setting configuration on a per-node basis. This allows overriding :ref:`cilium-config-configmap` for a node or set of nodes. It is managed by CiliumNodeConfig objects. This feature is useful for: - Gradually rolling out changes. - Selectively enabling features that require specific hardware: \* :ref:`XDP acceleration` \* :ref:`ipv6\_big\_tcp` CiliumNodeConfig objects ------------------------ A CiliumNodeConfig object allows for overriding ConfigMap / Agent arguments. It consists of a set of fields and a label selector. The label selector defines to which nodes the configuration applies. As is the standard with Kubernetes, an empty LabelSelector (e.g. ``{}``) selects all nodes. .. note:: Creating or modifying a CiliumNodeConfig will not cause changes to take effect until pods are deleted and re-created (or their node is restarted). Example: selective XDP enablement --------------------------------- To enable :ref:`XDP acceleration` only on nodes with necessary hardware, one would label the relevant nodes and override their configuration. .. code-block:: yaml apiVersion: cilium.io/v2 kind: CiliumNodeConfig metadata: namespace: kube-system name: enable-xdp spec: nodeSelector: matchLabels: io.cilium.xdp-offload: "true" defaults: bpf-lb-acceleration: native Example: KubeProxyReplacement Rollout ------------------------------------- To roll out :ref:`kube-proxy replacement ` in a gradual manner, you may also wish to use the CiliumNodeConfig feature. This will label all migrated nodes with ``io.cilium.migration/kube-proxy-replacement: true`` .. warning:: You must have installed Cilium with the Helm values ``k8sServiceHost`` and ``k8sServicePort``. Otherwise, Cilium will not be able to reach the Kubernetes APIServer after kube-proxy is uninstalled. You can apply these two values to a running cluster via ``helm upgrade``. #. Patch kube-proxy to only run on unmigrated nodes. .. code-block:: shell-session kubectl -n kube-system patch daemonset kube-proxy --patch '{"spec": {"template": {"spec": {"affinity": {"nodeAffinity": {"requiredDuringSchedulingIgnoredDuringExecution": {"nodeSelectorTerms": [{"matchExpressions": [{"key": "io.cilium.migration/kube-proxy-replacement", "operator": "NotIn", "values": ["true"]}]}]}}}}}}}' #. Configure Cilium to use kube-proxy replacement on migrated nodes .. code-block:: shell-session cat < | https://github.com/cilium/cilium/blob/main//Documentation/configuration/per-node-config.rst | main | cilium | [
0.01077566109597683,
0.03286318853497505,
-0.07807676494121552,
0.0018824589205905795,
0.03975209966301918,
-0.06644141674041748,
-0.044916409999132156,
0.028410974889993668,
-0.0013802782632410526,
-0.027600334957242012,
0.059242699295282364,
-0.040565766394138336,
0.06162550672888756,
-0... | 0.182943 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_gs\_dns: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Locking Down External Access with DNS-Based Policies \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction for using Cilium to enforce DNS-based security policies for Kubernetes pods. .. include:: gsg\_requirements.rst Deploy the Demo Application =========================== DNS-based policies are very useful for controlling access to services running outside the Kubernetes cluster. DNS acts as a persistent service identifier for both external services provided by AWS, Google, Twilio, Stripe, etc., and internal services such as database clusters running in private subnets outside Kubernetes. CIDR or IP-based policies are cumbersome and hard to maintain as the IPs associated with external services can change frequently. The Cilium DNS-based policies provide an easy mechanism to specify access control while Cilium manages the harder aspects of tracking DNS to IP mapping. In this guide we will learn about: - Controlling egress access to services outside the cluster using DNS-based policies - Using patterns (or wildcards) to whitelist a subset of DNS domains - Combining DNS, port and L7 rules for restricting access to external service In line with our Star Wars theme examples, we will use a simple scenario where the Empire's ``mediabot`` pods need access to GitHub for managing the Empire's git repositories. The pods shouldn't have access to any other external service. .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-sw-app.yaml $ kubectl wait pod/mediabot --for=condition=Ready $ kubectl get pods NAME READY STATUS RESTARTS AGE pod/mediabot 1/1 Running 0 14s Apply DNS Egress Policy ======================= The following Cilium network policy allows ``mediabot`` pods to only access ``api.github.com``. .. tabs:: .. group-tab:: Generic .. literalinclude:: ../../examples/kubernetes-dns/dns-matchname.yaml :language: yaml .. group-tab:: OpenShift .. literalinclude:: ../../examples/kubernetes-dns/dns-matchname-openshift.yaml :language: yaml .. note:: OpenShift users will need to modify the policies to match the namespace ``openshift-dns`` (instead of ``kube-system``), remove the match on the ``k8s:k8s-app=kube-dns`` label, and change the port to 5353. Let's take a closer look at the policy: \* The first egress section uses ``toFQDNs: matchName`` specification to allow egress to ``api.github.com``. The destination DNS should match exactly the name specified in the rule. The ``endpointSelector`` allows only pods with labels ``class: mediabot, org:empire`` to have the egress access. \* The second egress section (``toEndpoints``) allows ``mediabot`` pods to access ``kube-dns`` service. Note that ``rules: dns`` instructs Cilium to inspect and allow DNS lookups matching specified patterns. In this case, inspect and allow all DNS queries. Note that with this policy the ``mediabot`` doesn't have access to any internal cluster service other than ``kube-dns``. Refer to :ref:`Network Policy` to learn more about policies for controlling access to internal cluster services. Let's apply the policy: .. parsed-literal:: kubectl apply -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-matchname.yaml Testing the policy, we see that ``mediabot`` has access to ``api.github.com`` but doesn't have access to any other external service, e.g., ``support.github.com``. .. code-block:: shell-session $ kubectl exec mediabot -- curl -I -s https://api.github.com | head -1 HTTP/2 200 $ kubectl exec mediabot -- curl -I -s --max-time 5 https://support.github.com | head -1 curl: (28) Connection timed out after 5000 milliseconds command terminated with exit code 28 DNS Policies Using Patterns =========================== The above policy controlled DNS access based on exact match of the DNS domain name. Often, it is required to allow access to a subset of domains. Let's say, in the above example, ``mediabot`` pods need access to any GitHub sub-domain, e.g., the pattern ``\*.github.com``. We can achieve this easily by changing the ``toFQDN`` rule to use ``matchPattern`` instead of ``matchName``. .. literalinclude:: ../../examples/kubernetes-dns/dns-pattern.yaml :language: yaml .. parsed-literal:: kubectl apply | https://github.com/cilium/cilium/blob/main//Documentation/security/dns.rst | main | cilium | [
-0.07236611098051071,
0.01066063903272152,
-0.0708555206656456,
-0.0821755900979042,
-0.06148515269160271,
-0.10982102155685425,
-0.06913013756275177,
-0.021031849086284637,
0.025510651990771294,
-0.03317279741168022,
-0.009028870612382889,
-0.022619662806391716,
-0.006103378254920244,
-0.... | 0.148594 |
allow access to a subset of domains. Let's say, in the above example, ``mediabot`` pods need access to any GitHub sub-domain, e.g., the pattern ``\*.github.com``. We can achieve this easily by changing the ``toFQDN`` rule to use ``matchPattern`` instead of ``matchName``. .. literalinclude:: ../../examples/kubernetes-dns/dns-pattern.yaml :language: yaml .. parsed-literal:: kubectl apply -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-pattern.yaml Test that ``mediabot`` has access to multiple GitHub services for which the DNS matches the pattern ``\*.github.com``. It is important to note and test that this doesn't allow access to ``github.com`` because the ``\*.`` in the pattern requires one subdomain to be present in the DNS name. You can simply add more ``matchName`` and ``matchPattern`` clauses to extend the access. (See :ref:`DNS based` policies to learn more about specifying DNS rules using patterns and names.) .. code-block:: shell-session $ kubectl exec mediabot -- curl -I -s https://support.github.com | head -1 HTTP/1.1 200 OK $ kubectl exec mediabot -- curl -I -s https://gist.github.com | head -1 HTTP/1.1 302 Found $ kubectl exec mediabot -- curl -I -s --max-time 5 https://github.com | head -1 curl: (28) Connection timed out after 5000 milliseconds command terminated with exit code 28 Combining DNS, Port and L7 Rules ================================ The DNS-based policies can be combined with port (L4) and API (L7) rules to further restrict the access. In our example, we will restrict ``mediabot`` pods to access GitHub services only on ports ``443``. The ``toPorts`` section in the policy below achieves the port-based restrictions along with the DNS-based policies. .. literalinclude:: ../../examples/kubernetes-dns/dns-port.yaml :language: yaml .. parsed-literal:: kubectl apply -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-port.yaml Testing, the access to ``https://support.github.com`` on port ``443`` will succeed but the access to ``http://support.github.com`` on port ``80`` will be denied. .. code-block:: shell-session $ kubectl exec mediabot -- curl -I -s https://support.github.com | head -1 HTTP/1.1 200 OK $ kubectl exec mediabot -- curl -I -s --max-time 5 http://support.github.com | head -1 curl: (28) Connection timed out after 5001 milliseconds command terminated with exit code 28 Refer to :ref:`l4\_policy` and :ref:`l7\_policy` to learn more about Cilium L4 and L7 network policies. Clean-up ======== .. parsed-literal:: kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-sw-app.yaml kubectl delete cnp fqdn Limitations =========== By default, Cilium uses an in-agent DNS proxy for DNS policy enforcement. For high availability, consider using the :ref:`standalone\_dns\_proxy` (alpha feature). | https://github.com/cilium/cilium/blob/main//Documentation/security/dns.rst | main | cilium | [
-0.029559923335909843,
-0.04413173347711563,
0.03561205789446831,
-0.0985908955335617,
-0.024924246594309807,
-0.05337883532047272,
-0.007366642821580172,
-0.06632161140441895,
0.08700405806303024,
0.008520622737705708,
-0.04214642569422722,
-0.054062794893980026,
0.048617225140333176,
0.0... | 0.073869 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_host\_firewall: \*\*\*\*\*\*\*\*\*\*\*\*\* Host Firewall \*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction to Cilium's host firewall, to enforce security policies for Kubernetes nodes. .. admonition:: Video :class: attention You can also watch a video of Cilium's host firewall in action on `eCHO Episode 40: Cilium Host Firewall `\_\_. Enable the Host Firewall in Cilium ================================== .. include:: /installation/k8s-install-download-release.rst Deploy Cilium release via Helm: .. cilium-helm-install:: :namespace: kube-system :set: hostFirewall.enabled=true devices='{ethX,ethY}' The ``devices`` flag refers to the network devices Cilium is configured on, such as ``eth0``. If you omit this option, Cilium auto-detects what interfaces the host firewall applies to. The resulting interfaces are shown in the output of the ``cilium-dbg status`` command: .. code-block:: shell-session $ kubectl exec -n kube-system ds/cilium -- \ cilium-dbg status | grep 'Host firewall' At this point, the Cilium-managed nodes are ready to enforce network policies. Attach a Label to the Node ========================== In this guide, host policies only apply to nodes with the label ``node-access=ssh``. Therefore, you first need to attach this label to a node in the cluster: .. code-block:: shell-session $ export NODE\_NAME=k8s1 $ kubectl label node $NODE\_NAME node-access=ssh node/k8s1 labeled Enable Policy Audit Mode for the Host Endpoint ============================================== `HostPolicies` enforce access control over connectivity to and from nodes. Particular care must be taken to ensure that when host policies are imported, Cilium does not block access to the nodes or break the cluster's normal behavior (for example by blocking communication with ``kube-apiserver``). To avoid such issues, switch the host firewall in audit mode and validate the impact of host policies before enforcing them. .. warning:: When Policy Audit Mode is enabled, no network policy is enforced so this setting is not recommended for production deployment. .. note:: Audit mode does not persist across ``cilium-agent`` restarts. Once the agent is restarted, it will immediately \*\*enforce\*\* any existing host policies. Enable and check status for the Policy Audit Mode on the host endpoint for a given node with the following commands: .. code-block:: shell-session $ CILIUM\_NAMESPACE=kube-system $ CILIUM\_POD\_NAME=$(kubectl -n $CILIUM\_NAMESPACE get pods -l "k8s-app=cilium" -o jsonpath="{.items[?(@.spec.nodeName=='$NODE\_NAME')].metadata.name}") $ alias kexec="kubectl -n $CILIUM\_NAMESPACE exec $CILIUM\_POD\_NAME --" $ HOST\_EP\_ID=$(kexec cilium-dbg endpoint list -o jsonpath='{[?(@.status.identity.id==1)].id}') $ kexec cilium-dbg status | grep 'Host firewall' Host firewall: Enabled [eth0] $ kexec cilium-dbg endpoint config $HOST\_EP\_ID PolicyAuditMode=Enabled Endpoint 3353 configuration updated successfully $ kexec cilium-dbg endpoint config $HOST\_EP\_ID | grep PolicyAuditMode PolicyAuditMode : Enabled Apply a Host Network Policy =========================== :ref:`HostPolicies` match on node labels using a :ref:`NodeSelector` to identify the nodes to which the policies applies. They apply only to the host namespace, including host-networking pods. They don't apply to communications between pods or between pods and the outside of the cluster, except if those pods are host-networking pods. The following policy applies to all nodes with the ``node-access=ssh`` label. It allows communications from outside the cluster only for TCP/22 and for ICMP (ping) echo requests. All communications from the cluster to the hosts are allowed. .. literalinclude:: ../../examples/policies/host/demo-host-policy.yaml :language: yaml To apply this policy, run: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/policies/host/demo-host-policy.yaml ciliumclusterwidenetworkpolicy.cilium.io/demo-host-policy created The host is represented as a special endpoint, with label ``reserved:host``, in the output of command ``cilium-dbg endpoint list``. Use this command to inspect the status of host policies: .. code-block:: shell-session $ kexec cilium-dbg endpoint list ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 266 Disabled Disabled 104 k8s:io.cilium.k8s.policy.cluster=default f00d::a0b:0:0:ef4e 10.16.172.63 ready k8s:io.cilium.k8s.policy.serviceaccount=coredns k8s:io.kubernetes.pod.namespace=kube-system k8s:k8s-app=kube-dns 1687 Disabled (Audit) Disabled 1 k8s:node-access=ssh | https://github.com/cilium/cilium/blob/main//Documentation/security/host-firewall.rst | main | cilium | [
0.04385380446910858,
0.0714598149061203,
-0.053262073546648026,
-0.07470069825649261,
0.06345272809267044,
-0.04748811572790146,
-0.01201818510890007,
-0.04991938918828964,
0.0694640725851059,
0.004107477609068155,
0.015765486285090446,
-0.05269977077841759,
0.013974355533719063,
-0.027767... | 0.182203 |
endpoint list``. Use this command to inspect the status of host policies: .. code-block:: shell-session $ kexec cilium-dbg endpoint list ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 266 Disabled Disabled 104 k8s:io.cilium.k8s.policy.cluster=default f00d::a0b:0:0:ef4e 10.16.172.63 ready k8s:io.cilium.k8s.policy.serviceaccount=coredns k8s:io.kubernetes.pod.namespace=kube-system k8s:k8s-app=kube-dns 1687 Disabled (Audit) Disabled 1 k8s:node-access=ssh ready reserved:host 3362 Disabled Disabled 4 reserved:health f00d::a0b:0:0:49cf 10.16.87.66 ready In this example, one can observe that policy enforcement on the host endpoint is in audit mode for ingress traffic, and disabled for egress traffic. Adjust the Host Policy to Your Environment ========================================== As long as the host endpoint runs in audit mode, communications disallowed by the policy are not dropped. Nevertheless, they are reported by ``cilium-dbg monitor``, as ``action audit``. With these reports, the audit mode allows you to adjust the host policy to your environment in order to avoid unexpected connection breakages. .. code-block:: shell-session $ kexec cilium-dbg monitor -t policy-verdict --related-to $HOST\_EP\_ID Policy verdict log: flow 0x0 local EP ID 1687, remote ID 6, proto 1, ingress, action allow, match L3-Only, 192.168.60.12 -> 192.168.60.11 EchoRequest Policy verdict log: flow 0x0 local EP ID 1687, remote ID 6, proto 6, ingress, action allow, match L3-Only, 192.168.60.12:37278 -> 192.168.60.11:2379 tcp SYN Policy verdict log: flow 0x0 local EP ID 1687, remote ID 2, proto 6, ingress, action audit, match none, 10.0.2.2:47500 -> 10.0.2.15:6443 tcp SYN For details on deriving the network policies from the output of ``cilium monitor``, refer to `observe\_policy\_verdicts` and `create\_network\_policy` in the `policy\_verdicts` guide. Note that `Entities based` rules are convenient when combined with host policies, for example to allow communication to entire classes of destinations, such as all remotes nodes (``remote-node``) or the entire cluster (``cluster``). .. warning:: Make sure that none of the communications required to access the cluster or for the cluster to work properly are denied. Ensure they all appear as ``action allow`` before disabling the audit mode. .. \_disable\_policy\_audit\_mode: Disable Policy Audit Mode ========================= Once you are confident all required communications to the host from outside the cluster are allowed, disable the policy audit mode to enforce the host policy: .. code-block:: shell-session $ kexec cilium-dbg endpoint config $HOST\_EP\_ID PolicyAuditMode=Disabled Endpoint 3353 configuration updated successfully Ingress host policies should now appear as enforced: .. code-block:: shell-session $ kexec cilium-dbg endpoint list ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 266 Disabled Disabled 104 k8s:io.cilium.k8s.policy.cluster=default f00d::a0b:0:0:ef4e 10.16.172.63 ready k8s:io.cilium.k8s.policy.serviceaccount=coredns k8s:io.kubernetes.pod.namespace=kube-system k8s:k8s-app=kube-dns 1687 Enabled Disabled 1 k8s:node-access=ssh ready reserved:host 3362 Disabled Disabled 4 reserved:health f00d::a0b:0:0:49cf 10.16.87.66 ready Communications that are not explicitly allowed by the host policy are now dropped: .. code-block:: shell-session $ kexec cilium-dbg monitor -t policy-verdict --related-to $HOST\_EP\_ID Policy verdict log: flow 0x0 local EP ID 1687, remote ID 2, proto 6, ingress, action deny, match none, 10.0.2.2:49038 -> 10.0.2.15:21 tcp SYN Clean up ======== .. code-block:: shell-session $ kubectl delete ccnp demo-host-policy $ kubectl label node $NODE\_NAME node-access- Further Reading =============== Read the documentation on :ref:`HostPolicies` for additional details on how to use the policies. In particular, refer to the :ref:`Troubleshooting Host Policies ` subsection to understand how to debug issues with Host Policies, or to the section on :ref:`Host Policies known issues ` to understand the current limitations of the feature. Emergency Recovery ================== As host policies control access to the node, it is possible to create a policy that drops all access to nodes. In particular, if the Cilium agent loses access to the apiserver, it will not learn of any policy updates or deletes. This makes recovery complicated. If you have out-of-band access to the node(s), then it is possible to | https://github.com/cilium/cilium/blob/main//Documentation/security/host-firewall.rst | main | cilium | [
0.07495875656604767,
0.06669861078262329,
-0.02966560423374176,
-0.030221840366721153,
0.036736760288476944,
-0.02128114178776741,
0.02201811783015728,
-0.06826155632734299,
0.06558730453252792,
0.1025296226143837,
0.003913740627467632,
-0.08846402913331985,
-0.049238331615924835,
-0.04782... | 0.227345 |
is possible to create a policy that drops all access to nodes. In particular, if the Cilium agent loses access to the apiserver, it will not learn of any policy updates or deletes. This makes recovery complicated. If you have out-of-band access to the node(s), then it is possible to force-disable host policy enforcement and recover control. Start by deleting the offending host-firewall policy. Then, disable host policy enforcement manually on a node-by-node basis. Cilium Agent access ^^^^^^^^^^^^^^^^^^^ You will need to access the Cilium agent container. If ``kubelet`` still has network access, use ``kubectl exec``: .. code-block:: shell-session $ kubectl -n kube-system exec -ti $(CILIUM\_PODNAME) -- bash If this is unavailable, you will need ssh or console access to the node (e.g. via IPMI). Then, use ``crictl exec``: .. code-block:: shell-session $ CONTAINERID=$(crictl ps -q --label io.kubernetes.container.name=cilium-agent --label io.kubernetes.pod.namespace=kube-system) $ crictl exec -ti $CONTAINERID bash Disabling Host Policy enforcement ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ .. note:: This only disables host policy enforcement temporarily. Once ``cilium-agent`` is restarted, it will once again enforce host policies. Once you have access to the Cilium container, you can temporary disable host policy enforcement by enabling audit mode for the internal host endpoint. Audit mode converts policy drops in to a warning. .. code-block:: shell-session $ cilium-dbg endpoint config $(cilium-dbg endpoint get -l reserved:host -o 'jsonpath={$[0].id}') PolicyAuditMode=Enabled At this point, ``cilium-agent`` on the node will re-connect to the apiserver and synchronize policies. To re-enable host policy enforcement, either re-start the Cilium daemonset via ``kubectl rollout restart`` or manually: .. code-block:: shell-session $ cilium-dbg endpoint config $(cilium-dbg endpoint get -l reserved:host -o 'jsonpath={$[0].id}') PolicyAuditMode=Disabled | https://github.com/cilium/cilium/blob/main//Documentation/security/host-firewall.rst | main | cilium | [
0.043146517127752304,
0.08073480427265167,
-0.05628087371587753,
-0.03312227502465248,
-0.015167911536991596,
-0.03184480592608452,
-0.020411184057593346,
-0.017602279782295227,
0.036351606249809265,
0.039933837950229645,
-0.0007349600200541317,
-0.09732015430927277,
0.028034036979079247,
... | 0.158798 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Securing Elasticsearch \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction for using Cilium to enforce Elasticsearch-aware security policies. It is a detailed walk-through of getting a single-node Cilium environment running on your machine. It is designed to take 15-30 minutes. .. include:: gsg\_requirements.rst Deploy the Demo Application =========================== Following the Cilium tradition, we will use a Star Wars-inspired example. The Empire has a large scale Elasticsearch cluster which is used for storing a variety of data including: \* ``index: troop\_logs``: Stormtroopers performance logs collected from every outpost which are used to identify and eliminate weak performers! \* ``index: spaceship\_diagnostics``: Spaceships diagnostics data collected from every spaceship which is used for R&D and improvement of the spaceships. Every outpost has an Elasticsearch client service to upload the Stormtroopers logs. And every spaceship has a service to upload diagnostics. Similarly, the Empire headquarters has a service to search and analyze the troop logs and spaceship diagnostics data. Before we look into the security concerns, let's first create this application scenario in minikube. Deploy the app using command below, which will create \* An ``elasticsearch`` service with the selector label ``component:elasticsearch`` and a pod running Elasticsearch. \* Three Elasticsearch clients one each for ``empire-hq``, ``outpost`` and ``spaceship``. .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-es/es-sw-app.yaml serviceaccount "elasticsearch" created service "elasticsearch" created replicationcontroller "es" created role "elasticsearch" created rolebinding "elasticsearch" created pod "outpost" created pod "empire-hq" created pod "spaceship" created .. code-block:: shell-session $ kubectl get svc,pods NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE svc/elasticsearch NodePort 10.111.238.254 9200:30130/TCP,9300:31721/TCP 2d svc/etcd-cilium NodePort 10.98.67.60 32379:31079/TCP,32380:31080/TCP 9d svc/kubernetes ClusterIP 10.96.0.1 443/TCP 9d NAME READY STATUS RESTARTS AGE po/empire-hq 1/1 Running 0 2d po/es-g9qk2 1/1 Running 0 2d po/etcd-cilium-0 1/1 Running 0 9d po/outpost 1/1 Running 0 2d po/spaceship 1/1 Running 0 2d Security Risks for Elasticsearch Access ======================================= For Elasticsearch clusters the \*\*least privilege security\*\* challenge is to give clients access only to particular indices, and to limit the operations each client is allowed to perform on each index. In this example, the ``outpost`` Elasticsearch clients only need access to upload troop logs; and the ``empire-hq`` client only needs search access to both the indices. From the security perspective, the outposts are weak spots and susceptible to be captured by the rebels. Once compromised, the clients can be used to search and manipulate the critical data in Elasticsearch. We can simulate this attack, but first let's run the commands for legitimate behavior for all the client services. ``outpost`` client uploading troop logs .. code-block:: shell-session $ kubectl exec outpost -- python upload\_logs.py Uploading Stormtroopers Performance Logs created : {'\_index': 'troop\_logs', '\_type': 'log', '\_id': '1', '\_version': 1, 'result': 'created', '\_shards': {'total': 2, 'successful': 1, 'failed': 0}, 'created': True} ``spaceship`` uploading diagnostics .. code-block:: shell-session $ kubectl exec spaceship -- python upload\_diagnostics.py Uploading Spaceship Diagnostics created : {'\_index': 'spaceship\_diagnostics', '\_type': 'stats', '\_id': '1', '\_version': 1, 'result': 'created', '\_shards': {'total': 2, 'successful': 1, 'failed': 0}, 'created': True} ``empire-hq`` running search queries for logs and diagnostics .. code-block:: shell-session $ kubectl exec empire-hq -- python search.py Searching for Spaceship Diagnostics Got 1 Hits: {'\_index': 'spaceship\_diagnostics', '\_type': 'stats', '\_id': '1', '\_score': 1.0, \ '\_source': {'spaceshipid': '3459B78XNZTF', 'type': 'tiefighter', 'title': 'Engine Diagnostics', \ 'stats': '[CRITICAL] [ENGINE BURN @SPEED 5000 km/s] [CHANCE 80%]'}} Searching for Stormtroopers Performance Logs Got 1 Hits: {'\_index': 'troop\_logs', '\_type': 'log', '\_id': '1', '\_score': 1.0, \ '\_source': {'outpost': 'Endor', 'datetime': '33 ABY 4AM DST', 'title': 'Endor Corps 1: Morning Drill', \ 'notes': '5100 PRESENT; 15 ABSENT; | https://github.com/cilium/cilium/blob/main//Documentation/security/elasticsearch.rst | main | cilium | [
0.0006258433568291366,
0.055378999561071396,
-0.04775281995534897,
-0.020799897611141205,
0.04366708919405937,
-0.08848697692155838,
-0.06450550258159637,
0.021936338394880295,
-0.013401881791651249,
0.010068917647004128,
-0.015717841684818268,
-0.05729149654507637,
0.06503283977508545,
-0... | 0.179328 |
'title': 'Engine Diagnostics', \ 'stats': '[CRITICAL] [ENGINE BURN @SPEED 5000 km/s] [CHANCE 80%]'}} Searching for Stormtroopers Performance Logs Got 1 Hits: {'\_index': 'troop\_logs', '\_type': 'log', '\_id': '1', '\_score': 1.0, \ '\_source': {'outpost': 'Endor', 'datetime': '33 ABY 4AM DST', 'title': 'Endor Corps 1: Morning Drill', \ 'notes': '5100 PRESENT; 15 ABSENT; 130 CODE-RED BELOW PAR PERFORMANCE'}} Now imagine an outpost captured by the rebels. In the commands below, the rebels first search all the indices and then manipulate the diagnostics data from a compromised outpost. .. code-block:: shell-session $ kubectl exec outpost -- python search.py Searching for Spaceship Diagnostics Got 1 Hits: {'\_index': 'spaceship\_diagnostics', '\_type': 'stats', '\_id': '1', '\_score': 1.0, \ '\_source': {'spaceshipid': '3459B78XNZTF', 'type': 'tiefighter', 'title': 'Engine Diagnostics', \ 'stats': '[CRITICAL] [ENGINE BURN @SPEED 5000 km/s] [CHANCE 80%]'}} Searching for Stormtroopers Performance Logs Got 1 Hits: {'\_index': 'troop\_logs', '\_type': 'log', '\_id': '1', '\_score': 1.0, \ '\_source': {'outpost': 'Endor', 'datetime': '33 ABY 4AM DST', 'title': 'Endor Corps 1: Morning Drill', \ 'notes': '5100 PRESENT; 15 ABSENT; 130 CODE-RED BELOW PAR PERFORMANCE'}} Rebels manipulate spaceship diagnostics data so that the spaceship defects are not known to the empire-hq! (Hint: Rebels have changed the ``stats`` for the tiefighter spaceship, a change hard to detect but with adverse impact!) .. code-block:: shell-session $ kubectl exec outpost -- python update.py Uploading Spaceship Diagnostics {'\_index': 'spaceship\_diagnostics', '\_type': 'stats', '\_id': '1', '\_score': 1.0, \ '\_source': {'spaceshipid': '3459B78XNZTF', 'type': 'tiefighter', 'title': 'Engine Diagnostics', \ 'stats': '[OK] [ENGINE OK @SPEED 5000 km/s]'}} Securing Elasticsearch Using Cilium ==================================== .. image:: images/cilium\_es\_gsg\_topology.png :scale: 40 % Following the least privilege security principle, we want to the allow the following legitimate actions and nothing more: \* ``outpost`` service only has upload access to ``index: troop\_logs`` \* ``spaceship`` service only has upload access to ``index: spaceship\_diagnostics`` \* ``empire-hq`` service only has search access for both the indices Fortunately, the Empire DevOps team is using Cilium for their Kubernetes cluster. Cilium provides L7 visibility and security policies to control Elasticsearch API access. Cilium follows the \*\*white-list, least privilege model\*\* for security. That is to say, a \*CiliumNetworkPolicy\* contains a list of rules that define \*\*allowed requests\*\* and any request that does not match the rules is denied. In this example, the policy rules are defined for inbound traffic (i.e., "ingress") connections to the \*elasticsearch\* service. Note that endpoints selected as backend pods for the service are defined by the \*selector\* labels. \*Selector\* labels use the same concept as Kubernetes to define a service. In this example, label ``component: elasticsearch`` defines the pods that are part of the \*elasticsearch\* service in Kubernetes. In the policy file below, you will see the following rules for controlling the indices access and actions performed: \* ``fromEndpoints`` with labels ``app:spaceship`` only ``HTTP`` ``PUT`` is allowed on paths matching regex ``^/spaceship\_diagnostics/stats/.\*$`` \* ``fromEndpoints`` with labels ``app:outpost`` only ``HTTP`` ``PUT`` is allowed on paths matching regex ``^/troop\_logs/log/.\*$`` \* ``fromEndpoints`` with labels ``app:empire`` only ``HTTP`` ``GET`` is allowed on paths matching regex ``^/spaceship\_diagnostics/\_search/??.\*$`` and ``^/troop\_logs/search/??.\*$`` .. literalinclude:: ../../examples/kubernetes-es/es-sw-policy.yaml :language: yaml Apply this Elasticsearch-aware network security policy using ``kubectl``: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-es/es-sw-policy.yaml ciliumnetworkpolicy "secure-empire-elasticsearch" created Let's test the security policies. Firstly, the search access is blocked for both outpost and spaceship. So from a compromised outpost, rebels will not be able to search and obtain knowledge about troops and spaceship diagnostics. Secondly, the outpost clients don't have access to create or update the ``index: spaceship\_diagnostics``. .. code-block:: shell-session $ kubectl exec outpost -- python search.py GET http://elasticsearch:9200/spaceship\_diagnostics/\_search [status:403 request:0.008s] ... ... elasticsearch.exceptions.AuthorizationException: TransportError(403, 'Access denied\r\n') command terminated with exit code 1 .. code-block:: shell-session $ kubectl exec outpost -- | https://github.com/cilium/cilium/blob/main//Documentation/security/elasticsearch.rst | main | cilium | [
0.010850842110812664,
0.06034308671951294,
-0.022837592288851738,
0.039641425013542175,
0.057596515864133835,
-0.05435784533619881,
0.027199719101190567,
0.014008739963173866,
0.04719487950205803,
0.022738123312592506,
0.0474558100104332,
-0.05543138459324837,
0.0238949041813612,
-0.032500... | 0.142311 |
spaceship diagnostics. Secondly, the outpost clients don't have access to create or update the ``index: spaceship\_diagnostics``. .. code-block:: shell-session $ kubectl exec outpost -- python search.py GET http://elasticsearch:9200/spaceship\_diagnostics/\_search [status:403 request:0.008s] ... ... elasticsearch.exceptions.AuthorizationException: TransportError(403, 'Access denied\r\n') command terminated with exit code 1 .. code-block:: shell-session $ kubectl exec outpost -- python update.py PUT http://elasticsearch:9200/spaceship\_diagnostics/stats/1 [status:403 request:0.006s] ... ... elasticsearch.exceptions.AuthorizationException: TransportError(403, 'Access denied\r\n') command terminated with exit code 1 We can re-run any of the below commands to show that the security policy still allows all legitimate requests (i.e., no 403 errors are returned). .. code-block:: shell-session $ kubectl exec outpost -- python upload\_logs.py ... $ kubectl exec spaceship -- python upload\_diagnostics.py ... $ kubectl exec empire-hq -- python search.py ... Clean Up ======== You have now installed Cilium, deployed a demo app, and finally deployed & tested Elasticsearch-aware network security policies. To clean up, run: .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-es/es-sw-app.yaml $ kubectl delete cnp secure-empire-elasticsearch | https://github.com/cilium/cilium/blob/main//Documentation/security/elasticsearch.rst | main | cilium | [
0.05937686562538147,
0.05467497184872627,
-0.019312679767608643,
-0.002743070712313056,
-0.013596885837614536,
-0.055692244321107864,
-0.07699821889400482,
-0.0005754122976213694,
0.05003131181001663,
0.06406761705875397,
-0.010233674198389053,
-0.0447072871029377,
-0.004421275109052658,
0... | 0.131746 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_gs\_http: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Identity-Aware and HTTP-Aware Policy Enforcement \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* .. include:: gsg\_requirements.rst When you have Cilium installed, follow the :ref:`starwars\_demo` tutorial to walk you through Identity-Aware and HTTP-Aware Policy Enforcement. | https://github.com/cilium/cilium/blob/main//Documentation/security/http.rst | main | cilium | [
-0.0710919126868248,
0.06790466606616974,
-0.0955081656575203,
-0.029901113361120224,
0.03050639107823372,
-0.08984628319740295,
-0.023529822006821632,
-0.012066110037267208,
0.022885248064994812,
0.015297790989279747,
0.04091818258166313,
-0.040596287697553635,
0.046231675893068314,
-0.03... | 0.175495 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_standalone\_dns\_proxy: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Standalone DNS Proxy (alpha) \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* .. include:: ../alpha.rst The Standalone DNS Proxy is an independent component that runs as a separate DaemonSet in the cluster, providing DNS proxying capabilities independent of the Cilium agent. The in-agent proxy runs alongside the Standalone DNS Proxy. The load of DNS request is shared between both proxies. Overview ======== The Standalone DNS Proxy communicates with the Cilium agent via gRPC to: 1. Receive DNS policy rules from the agent 2. Report DNS query results for policy enforcement to the agent Configuration ============= To enable the Standalone DNS Proxy, set the following Helm values: .. code-block:: yaml # Enable L7 proxy and configure DNS proxy port l7Proxy: true dnsProxy: proxyPort: 10094 # Must be non-zero when using standalone DNS proxy, choosing 10094 as example # Enable standalone DNS proxy standaloneDnsProxy: enabled: true serverPort: 10095 # Must be non-zero when using standalone DNS proxy, choosing 10095 as example .. important:: The standalone DNS proxy uses the agent's DNS configuration to ensure consistency: \* ``dnsProxy.proxyPort`` must be explicitly set to a non-zero value (e.g., 10094) \* ``dnsProxy.enableDnsCompression`` and other DNS settings will automatically use the same defaults as the agent (default: true) The Helm chart will fail validation if ``proxyPort`` is not configured correctly. Testing the Standalone DNS Proxy ================================= This section provides steps to test the Standalone DNS Proxy in a development environment. To test the standalone DNS proxy feature, you need to build container images from source. The following instructions guide you through building and deploying the standalone DNS proxy in a local development environment using kind (Kubernetes in Docker). Building and Deploying ---------------------- #. \*\*Build the Standalone DNS Proxy Image\*\* Build the standalone DNS proxy container image from source: .. code-block:: shell-session $ make docker-standalone-dns-proxy-image This creates a local image ``quay.io/cilium/standalone-dns-proxy:latest``. #. \*\*Set Up kind Cluster with Cilium\*\* Create a kind cluster and build/install Cilium from source: .. code-block:: shell-session $ make kind && make kind-image && make kind-install-cilium #. \*\*Load the Standalone DNS Proxy Image into kind\*\* Load the standalone DNS proxy image you built into the kind cluster: .. code-block:: shell-session $ kind load docker-image quay.io/cilium/standalone-dns-proxy:latest #. \*\*Upgrade Cilium to Enable Standalone DNS Proxy\*\* Enable the standalone DNS proxy and configure it to work with the Cilium agent: .. code-block:: shell-session $ cilium upgrade \ --chart-directory='./install/kubernetes/cilium' \ --set='l7Proxy=true' \ --set='dnsProxy.proxyPort=10094' \ --helm-set='standaloneDnsProxy.enabled=true' \ --helm-set='standaloneDnsProxy.image.repository=quay.io/cilium/standalone-dns-proxy' \ --helm-set='standaloneDnsProxy.image.tag=latest' \ --helm-set='standaloneDnsProxy.image.useDigest=false' \ --helm-set='standaloneDnsProxy.image.pullPolicy=Never' The configuration flags in this example ensure the standalone proxy is operational by applying the following configurations: \* ``dnsProxy.proxyPort=10094`` sets the DNS proxy port used by both the agent and standalone DNS proxy \* ``l7Proxy=true`` enables L7 proxy support required for DNS policy enforcement \* The standalone DNS proxy automatically inherits DNS settings from the agent configuration \* ``image.tag=latest`` and ``image.pullPolicy=Never`` are used to reference the locally-built image #. \*\*Restart Cilium Agent\*\* Restart the Cilium agent to apply the configuration changes: .. code-block:: shell-session $ kubectl rollout restart ds -n kube-system cilium #. \*\*Verify Deployment\*\* Check that the standalone DNS proxy pods are running: .. code-block:: shell-session $ kubectl -n kube-system get pods -l k8s-app=standalone-dns-proxy NAME READY STATUS RESTARTS AGE standalone-dns-proxy-xxxxx 1/1 Running 0 1m #. \*\*Apply DNS Policy\*\* Apply a policy that allows pods with the label ``org: alliance`` to query specific domains (``cilium.io`` and its subdomains) and blocks all other queries: .. literalinclude:: ../../examples/policies/l7/dns/dns.yaml :language: yaml .. code-block:: shell-session $ kubectl apply -f examples/policies/l7/dns/dns.yaml #. \*\*Deploy Test Pod\*\* Create a test pod | https://github.com/cilium/cilium/blob/main//Documentation/security/standalone-dns-proxy.rst | main | cilium | [
-0.05569147691130638,
-0.044896796345710754,
-0.07918699085712433,
-0.061037566512823105,
-0.0575522817671299,
-0.052200186997652054,
-0.06506212055683136,
-0.02225487492978573,
0.012202208861708641,
-0.044258926063776016,
0.009435399435460567,
-0.014575833454728127,
-0.005848078988492489,
... | 0.17145 |
1m #. \*\*Apply DNS Policy\*\* Apply a policy that allows pods with the label ``org: alliance`` to query specific domains (``cilium.io`` and its subdomains) and blocks all other queries: .. literalinclude:: ../../examples/policies/l7/dns/dns.yaml :language: yaml .. code-block:: shell-session $ kubectl apply -f examples/policies/l7/dns/dns.yaml #. \*\*Deploy Test Pod\*\* Create a test pod with the matching label: .. code-block:: shell-session $ kubectl run test-pod --image=nicolaka/netshoot --labels="org=alliance" -- sleep 3600 #. \*\*Verify DNS Policy Enforcement\*\* Test that DNS queries are being intercepted and resolved for allowed domains: .. code-block:: shell-session $ kubectl exec test-pod -- nslookup cilium.io. Queries to ``cilium.io`` should succeed. Now test that queries to non-allowed domains are refused: .. code-block:: shell-session $ kubectl exec test-pod -- nslookup example.com. Server: 10.96.0.10 Address: 10.96.0.10#53 \*\* server can't find example.com: REFUSED The query to ``example.com`` is refused because it's not in the allowed DNS policy rules. #. \*\*Test Standalone Proxy Resilience\*\* Verify that the standalone DNS proxy continues to work even when the Cilium agent is down: .. code-block:: shell-session # Intentionally break the Cilium agent by using a non-existent image $ kubectl set image -n kube-system ds/cilium cilium-agent=quay.io/cilium/cilium:non-existent-image # Wait for agent pods to enter ImagePullBackOff state $ kubectl wait --for=condition=Ready=false pod -n kube-system -l k8s-app=cilium --timeout=60s # DNS queries for allowed domains should still work via standalone DNS proxy $ kubectl exec test-pod -- nslookup cilium.io # Verify that policy enforcement still works - non-allowed domains are still refused $ kubectl exec test-pod -- nslookup example.com Server: 10.96.0.10 Address: 10.96.0.10#53 \*\* server can't find example.com: REFUSED Both queries demonstrate that the standalone DNS proxy continues to enforce DNS policies independently, allowing ``cilium.io`` and refusing ``example.com``, even when the agent is unavailable. #. \*\*Restore Cilium Agent\*\* Restore the Cilium agent to normal operation: .. code-block:: shell-session $ kubectl rollout undo ds/cilium -n kube-system For more information on DNS policies, see :ref:`DNS based`. Limitations =========== The Standalone DNS Proxy alpha release has the following known limitations: \* The standalone DNS proxy uses the agent's DNS configuration. The ``dnsProxy.proxyPort`` must be explicitly set to a non-zero value when the standalone DNS proxy is enabled (it does not automatically select a free port). \* Metrics related to DNS are not supported yet. The metrics are currently only available from the in-agent DNS proxy. \* Standalone DNS proxy depends on Cilium agent to read DNS policies, enforce them and communicate via gRPC. If there are connectivity issues between the proxy and agent, DNS policy enforcement may be affected. \* While the standalone DNS proxy can continue to proxy DNS requests when the agent is down, it cannot allocate new identities for domains that haven't been observed before. If an endpoint looks up a new domain (one that hasn't been cached) while the agent is unavailable, the resulting traffic will be dropped because no security identity can be allocated. Only DNS lookups for previously observed domains (with cached identities) will work during agent downtime. Troubleshooting =============== Validation Errors ----------------- If the Helm chart fails with a validation error about ``dnsProxy.proxyPort``:: Error: standaloneDnsProxy requires dnsProxy.proxyPort to be set to a non-zero value (e.g., 10094) This means you need to explicitly configure the DNS proxy port in your values: .. code-block:: yaml dnsProxy: proxyPort: 10094 # Must be non-zero Port Configuration Verification ------------------------------- To verify that the DNS proxy port is correctly configured: .. code-block:: shell-session $ kubectl -n kube-system get configmap cilium-config -o yaml | grep 'tofqdns-proxy-port' $ kubectl -n kube-system get configmap standalone-dns-proxy-config -o yaml | grep 'tofqdns-proxy-port' Both ConfigMaps should show the same value from ``dnsProxy.proxyPort``. gRPC Communication Issues ------------------------- If there are communication issues between the | https://github.com/cilium/cilium/blob/main//Documentation/security/standalone-dns-proxy.rst | main | cilium | [
0.01168232411146164,
0.016173554584383965,
-0.0037241920363157988,
-0.07622082531452179,
-0.06851118803024292,
-0.07692359387874603,
-0.022724229842424393,
-0.08622345328330994,
0.06254410743713379,
0.04483296349644661,
-0.013178396038711071,
-0.09583081305027008,
-0.028747303411364555,
-0... | 0.144339 |
is correctly configured: .. code-block:: shell-session $ kubectl -n kube-system get configmap cilium-config -o yaml | grep 'tofqdns-proxy-port' $ kubectl -n kube-system get configmap standalone-dns-proxy-config -o yaml | grep 'tofqdns-proxy-port' Both ConfigMaps should show the same value from ``dnsProxy.proxyPort``. gRPC Communication Issues ------------------------- If there are communication issues between the proxy and agent, review agent logs for connection errors: .. code-block:: shell-session $ kubectl -n kube-system logs -l k8s-app=cilium --tail=100 | grep -i "fqdn.sdp-grpc-server" API Reference ============= For detailed API documentation, see :ref:`sdpapi\_ref`. Further Reading =============== \* :ref:`DNS based` \* :ref:`DNS Proxy` | https://github.com/cilium/cilium/blob/main//Documentation/security/standalone-dns-proxy.rst | main | cilium | [
-0.012698957696557045,
-0.007663814816623926,
-0.03619479015469551,
-0.07514207810163498,
-0.10840550810098648,
-0.03620848432183266,
0.002222121926024556,
-0.028461769223213196,
0.060523342341184616,
0.05589468032121658,
-0.015432722866535187,
-0.07851628214120865,
-0.02455184981226921,
-... | 0.08103 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io \*\*\*\*\*\*\*\*\*\*\*\*\* Securing gRPC \*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction to using Cilium to enforce gRPC-aware security policies. It is a detailed walk-through of getting a single-node Cilium environment running on your machine. It is designed to take 15-30 minutes. .. include:: gsg\_requirements.rst It is important for this demo that ``kube-dns`` is working correctly. To know the status of ``kube-dns`` you can run the following command: .. code-block:: shell-session $ kubectl get deployment kube-dns -n kube-system NAME DESIRED CURRENT UP-TO-DATE AVAILABLE AGE kube-dns 1 1 1 1 13h Where at least one pod should be available. Deploy the Demo Application =========================== Now that we have Cilium deployed and ``kube-dns`` operating correctly we can deploy our demo gRPC application. Since our first demo of Cilium + HTTP-aware security policies was Star Wars-themed, we decided to do the same for gRPC. While the `HTTP-aware Cilium Star Wars demo `\_ showed how the Galactic Empire used HTTP-aware security policies to protect the Death Star from the Rebel Alliance, this gRPC demo shows how the lack of gRPC-aware security policies allowed Leia, Chewbacca, Lando, C-3PO, and R2-D2 to escape from Cloud City, which had been overtaken by empire forces. `gRPC `\_ is a high-performance RPC framework built on top of the `protobuf `\_ serialization/deserialization library popularized by Google. There are gRPC bindings for many programming languages, and the efficiency of the protobuf parsing as well as advantages from leveraging HTTP 2 as a transport make it a popular RPC framework for those building new microservices from scratch. For those unfamiliar with the details of the movie, Leia and the other rebels are fleeing storm troopers and trying to reach the space port platform where the Millennium Falcon is parked, so they can fly out of Cloud City. However, the door to the platform is closed, and the access code has been changed. However, R2-D2 is able to access the Cloud City computer system via a public terminal, and disable this security, opening the door and letting the Rebels reach the Millennium Falcon just in time to escape. .. image:: images/cilium\_grpc\_gsg\_r2d2\_terminal.png In our example, Cloud City's internal computer system is built as a set of gRPC-based microservices (who knew that gRPC was actually invented a long time ago, in a galaxy far, far away?). With gRPC, each service is defined using a language independent protocol buffer definition. Here is the definition for the system used to manage doors within Cloud City: .. code-block:: java package cloudcity; // The door manager service definition. service DoorManager { // Get human readable name of door. rpc GetName(DoorRequest) returns (DoorNameReply) {} // Find the location of this door. rpc GetLocation (DoorRequest) returns (DoorLocationReply) {} // Find out whether door is open or closed rpc GetStatus(DoorRequest) returns (DoorStatusReply) {} // Request maintenance on the door rpc RequestMaintenance(DoorMaintRequest) returns (DoorActionReply) {} // Set Access Code to Open / Lock the door rpc SetAccessCode(DoorAccessCodeRequest) returns (DoorActionReply) {} } To keep the setup small, we will just launch two pods to represent this setup: - \*\*cc-door-mgr\*\*: A single pod running the gRPC door manager service with label ``app=cc-door-mgr``. - \*\*terminal-87\*\*: One of the public network access terminals scattered across Cloud City. R2-D2 plugs into terminal-87 as the rebels are desperately trying to escape. This terminal uses the gRPC client code to communicate with the door management services with label ``app=public-terminal``. .. image:: images/cilium\_grpc\_gsg\_topology.png The file ``cc-door-app.yaml`` contains a Kubernetes Deployment for the door manager service, a Kubernetes Pod | https://github.com/cilium/cilium/blob/main//Documentation/security/grpc.rst | main | cilium | [
0.002793343970552087,
0.00487463129684329,
-0.051869045943021774,
-0.0700882077217102,
-0.02421206794679165,
-0.07918273657560349,
-0.0749259740114212,
-0.012810954824090004,
0.05240652710199356,
-0.005952463950961828,
0.030716942623257637,
-0.08152509480714798,
-0.00669920863583684,
-0.05... | 0.138489 |
across Cloud City. R2-D2 plugs into terminal-87 as the rebels are desperately trying to escape. This terminal uses the gRPC client code to communicate with the door management services with label ``app=public-terminal``. .. image:: images/cilium\_grpc\_gsg\_topology.png The file ``cc-door-app.yaml`` contains a Kubernetes Deployment for the door manager service, a Kubernetes Pod representing ``terminal-87``, and a Kubernetes Service for the door manager services. To deploy this example app, run: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-grpc/cc-door-app.yaml deployment "cc-door-mgr" created service "cc-door-server" created pod "terminal-87" created Kubernetes will deploy the pods and service in the background. Running ``kubectl get svc,pods`` will inform you about the progress of the operation. Each pod will go through several states until it reaches ``Running`` at which point the setup is ready. .. code-block:: shell-session $ kubectl get pods,svc NAME READY STATUS RESTARTS AGE po/cc-door-mgr-3590146619-cv4jn 1/1 Running 0 1m po/terminal-87 1/1 Running 0 1m NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE svc/cc-door-server 10.0.0.72 50051/TCP 1m svc/kubernetes 10.0.0.1 443/TCP 6m Test Access Between gRPC Client and Server ========================================== First, let's confirm that the public terminal can properly act as a client to the door service. We can test this by running a Python gRPC client for the door service that exists in the \*terminal-87\* container. We'll invoke the 'cc\_door\_client' with the name of the gRPC method to call, and any parameters (in this case, the door-id): .. code-block:: shell-session $ kubectl exec terminal-87 -- python3 /cloudcity/cc\_door\_client.py GetName 1 Door name is: Spaceport Door #1 $ kubectl exec terminal-87 -- python3 /cloudcity/cc\_door\_client.py GetLocation 1 Door location is lat = 10.222200393676758 long = 68.87879943847656 Exposing this information to public terminals seems quite useful, as it helps travelers new to Cloud City identify and locate different doors. But recall that the door service also exposes several other methods, including ``SetAccessCode``. If access to the door manager service is protected only using traditional IP and port-based firewalling, the TCP port of the service (50051 in this example) will be wide open to allow legitimate calls like ``GetName`` and ``GetLocation``, which also leave more sensitive calls like ``SetAccessCode`` exposed as well. It is this mismatch between the course granularity of traditional firewalls and the fine-grained nature of gRPC calls that R2-D2 exploited to override the security and help the rebels escape. To see this, run: .. code-block:: shell-session $ kubectl exec terminal-87 -- python3 /cloudcity/cc\_door\_client.py SetAccessCode 1 999 Successfully set AccessCode to 999 Securing Access to a gRPC Service with Cilium ============================================= Once the legitimate owners of Cloud City recover the city from the empire, how can they use Cilium to plug this key security hole and block requests to ``SetAccessCode`` and ``GetStatus`` while still allowing ``GetName``, ``GetLocation``, and ``RequestMaintenance``? .. image:: images/cilium\_grpc\_gsg\_policy.png Since gRPC build on top of HTTP, this can be achieved easily by understanding how a gRPC call is mapped to an HTTP URL, and then applying a Cilium HTTP-aware filter to allow public terminals to only invoke a subset of all the total gRPC methods available on the door service. Each gRPC method is mapped to an HTTP POST call to a URL of the form ``/cloudcity.DoorManager/``. As a result, the following \*CiliumNetworkPolicy\* rule limits access of pods with label ``app=public-terminal`` to only invoke ``GetName``, ``GetLocation``, and ``RequestMaintenance`` on the door service, identified by label ``app=cc-door-mgr``: .. literalinclude:: ../../examples/kubernetes-grpc/cc-door-ingress-security.yaml :language: yaml :emphasize-lines: 9,13,21 A \*CiliumNetworkPolicy\* contains a list of rules that define allowed requests, meaning that requests that do not match any rules (e.g., ``SetAccessCode``) are denied as invalid. The above rule applies to inbound (i.e., "ingress") connections to ``cc-door-mgr pods`` (as indicated by ``app: cc-door-mgr`` in the "endpointSelector" section). The | https://github.com/cilium/cilium/blob/main//Documentation/security/grpc.rst | main | cilium | [
0.042416900396347046,
-0.05754779279232025,
-0.024747150018811226,
-0.04381299018859863,
-0.0013506164541468024,
-0.04599256440997124,
-0.01127244345843792,
-0.006574172060936689,
0.06858434528112411,
0.09768624603748322,
0.03582334518432617,
-0.03350532054901123,
0.006969935726374388,
-0.... | 0.086375 |
:emphasize-lines: 9,13,21 A \*CiliumNetworkPolicy\* contains a list of rules that define allowed requests, meaning that requests that do not match any rules (e.g., ``SetAccessCode``) are denied as invalid. The above rule applies to inbound (i.e., "ingress") connections to ``cc-door-mgr pods`` (as indicated by ``app: cc-door-mgr`` in the "endpointSelector" section). The rule will apply to connections from pods with label ``app: public-terminal`` as indicated by the "fromEndpoints" section. The rule explicitly matches gRPC connections destined to TCP 50051, and white-lists specifically the permitted URLs. Apply this gRPC-aware network security policy using ``kubectl`` in the main window: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-grpc/cc-door-ingress-security.yaml After this security policy is in place, access to the innocuous calls like ``GetLocation`` still works as intended: .. code-block:: shell-session $ kubectl exec terminal-87 -- python3 /cloudcity/cc\_door\_client.py GetLocation 1 Door location is lat = 10.222200393676758 long = 68.87879943847656 However, if we then again try to invoke ``SetAccessCode``, it is denied: .. code-block:: shell-session $ kubectl exec terminal-87 -- python3 /cloudcity/cc\_door\_client.py SetAccessCode 1 999 Traceback (most recent call last): File "/cloudcity/cc\_door\_client.py", line 71, in run() File "/cloudcity/cc\_door\_client.py", line 52, in run response = stub.SetAccessCode(cloudcity\_pb2.DoorAccessCodeRequest( File "/usr/local/lib/python3.8/dist-packages/grpc/\_channel.py", line 826, in \_\_call\_\_ return \_end\_unary\_response\_blocking(state, call, False, None) File "/usr/local/lib/python3.8/dist-packages/grpc/\_channel.py", line 729, in \_end\_unary\_response\_blocking raise \_InactiveRpcError(state) grpc.\_channel.\_InactiveRpcError: <\_InactiveRpcError of RPC that terminated with: status = StatusCode.PERMISSION\_DENIED details = "Access denied" debug\_error\_string = "{"created":"@1748342370.953859451","description":"Error received from peer ipv4:10.96.86.105:50051","file":"src/core/lib/surface/call.cc","file\_line":1055,"grpc\_message":"Access denied\r\n","grpc\_status":7}" This is now blocked, thanks to the Cilium network policy. And notice that unlike a traditional firewall which would just drop packets in a way indistinguishable from a network failure, because Cilium operates at the API-layer, it can explicitly reply with a custom gRPC status code 7 PERMISSION\_DENIED, indicating that the request was intentionally denied for security reasons. Thank goodness that the empire IT staff hadn't had time to deploy Cilium on Cloud City's internal network prior to the escape attempt, or things might have turned out quite differently for Leia and the other Rebels! Clean-Up ======== You have now installed Cilium, deployed a demo app, and tested L7 gRPC-aware network security policies. To clean-up, run: .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-grpc/cc-door-app.yaml $ kubectl delete cnp rule1 After this, you can re-run the tutorial from Step 1. | https://github.com/cilium/cilium/blob/main//Documentation/security/grpc.rst | main | cilium | [
0.011636133305728436,
0.032658420503139496,
-0.013538407161831856,
-0.038303252309560776,
-0.025490975007414818,
-0.03831407055258751,
0.011779882945120335,
-0.022150373086333275,
0.008565192110836506,
0.008856605738401413,
-0.013433657586574554,
-0.04623762145638466,
0.016999324783682823,
... | 0.165983 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_gs\_tls\_inspection: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Inspecting TLS Encrypted Connections with Cilium \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction for how network security teams can use Cilium to transparently inspect TLS-encrypted connections. This TLS-aware inspection allows Cilium API-aware visibility and policy to function even for connections where client to server communication is protected by TLS, such as when a client accesses the API service via HTTPS. This capability is similar to what is possible to traditional hardware firewalls, but is implemented entirely in software on the Kubernetes worker node, and is policy driven, allowing inspection to target only selected network connectivity. This type of visibility is extremely valuable to be able to monitor how external API services are being used, for example, understanding which S3 buckets are being accessed by an given application. .. include:: gsg\_requirements.rst A Brief Overview of the TLS Certificate Model ============================================= TLS is a protocol that "wraps" other protocols like HTTP and ensures that communication between client and server has confidentiality (no one can read the data except the intended recipient), integrity (recipient can confirm that the data has not been modified in transit), and authentication (sender can confirm that it is talking with the intended destination, not an impostor). We will provide a highly simplified overview of TLS in this document, but for full details, please see ``\_ . From an authentication perspective, the TLS model relies on a "Certificate Authority" (CA) which is an entity that is trusted to create proof that a given network service (e.g., www.cilium.io) is who they say they are. The goal is to prevents a malicious party in the network between the client and the server from intercepting the traffic and pretending to be the destination server. In the case of "friendly interception" for network security monitoring, Cilium uses a model similar to traditional firewalls with TLS inspection capabilities: the network security team creates their own "internal certificate authority" that can be used to create alternative certificates for external destinations. This model requires each client workload to also trust this new certificate, otherwise the client's TLS library will reject the connection as invalid. In this model, the network firewall uses the certificate signed by the internal CA to act like the destination service and terminate the TLS connection. This allows the firewall to inspect and even modify the application layer data, and then initiate another TLS connect to the actual destination service. The CA model within TLS is based on cryptographic keys and certificates. Realizing the above model requires four primary steps: 1) Create an internal certificate authority by generating a CA private key and CA certificate. 2) For any destination where TLS inspection is desired (e.g., httpbin.org in the example below), generate a private key and certificate signing request with a common name that matches the destination DNS name. 3) Use the CA private key to create a signed certificate. 4) Ensure that all clients where TLS inspection is have the CA certificate installed so that they will trust all certificates signed by that CA. 5) Given that Cilium will be terminating the initial TLS connection from the client and creating a new TLS connection to the destination, Cilium must be told the set of CAs that it should trust when validating the new TLS connection to the destination service. .. note:: In a non-demo environment it is EXTREMELY important that you keep the above private keys safe, as anyone with access to this private key | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.08302503824234009,
0.059587642550468445,
-0.07575305551290512,
-0.040874868631362915,
-0.004146907012909651,
-0.1120092049241066,
-0.03223421424627304,
-0.015262660570442677,
0.0678500384092331,
-0.030095847323536873,
0.0430421344935894,
-0.04721168428659439,
0.02786884270608425,
-0.025... | 0.137708 |
the destination, Cilium must be told the set of CAs that it should trust when validating the new TLS connection to the destination service. .. note:: In a non-demo environment it is EXTREMELY important that you keep the above private keys safe, as anyone with access to this private key will be able to inspect TLS-encrypted traffic (certificates on the other hand are public information, and are not at all sensitive). In the guide below, the CA private key does not need to be provided to Cilium at all (it is used only to create certificates, which can be done offline) and private keys for individual destination services are stored as Kubernetes secrets. These secrets should be stored in a namespace where they can be accessed by Cilium, but not general purpose workloads. How TLS Inspection works ======================== All TLS inspection relies on terminating the originating connection with a certificate that will be accepted, then originating a new TLS connection using a client certificate if necessary. Because of this, the Network Policy requires configuring a ``terminatingTLS`` and optionally an ``originatingTLS`` stanza. When the Network Policy contains these details, then Cilium will redirect TLS connections to Envoy, and allow connections that complete a TLS handshake and pass the configured Network Policy. One of the most important parts of the configuration for this is how the certificates get to Envoy. In the current version, Cilium has two options, NPDS (the original) and SDS (the new and better version). Network Policy Discovery Service (NPDS) --------------------------------------- In this version, certificates and keys are sent inline as Base64 encoded text in dedicated fields in the Cilium-owned Network Policy Discovery Service. This had the advantage that it was straightforward to build, but does come with a big disadvantage: Each Network Policy rule that does TLS Interception keeps its own copy of each secret inline in the NPDS config in Envoy. So, if (as is likely for a larger installation), you have the same secret reused multiple times (for example if you generate one certificate that will terminate for many SANs, but you have multiple rules using that certificate, or you include a valid root certificate bundle in the ``originatingTLS`` config), then multiple copies of the certificate will be stored in Envoy's memory. This memory use can really add up in a large installation. It also means that we don't benefit from work that has been done to protect secrets when they are sent using Secret Discovery Service (. Secret Discovery Service (SDS) ------------------------------ Both of the above reasons are why Envoy supports SDS for Network Policy secrets as of Cilium 1.17. In this configuration, Cilium reads relevant Secrets from a configured secrets namespace, and exposes those secrets to Envoy using the core Envoy SDS API. Those secrets are then referenced in the NPDS config that's sent to Envoy to configure the Network Policy filter there by name, rather than being included directly as Base64 encoded text. This means that Envoy looks up the SDS secrets for NPDS in the same way as it does the secrets for Ingress or Gateway API config. This method also allows Envoy to deduplicate the storage of the secrets, since they are essentially being passed by reference instead of being passed by value. Because of these advantages over the older NPDS method, SDS is the default for new Cilium installations as at Cilium 1.17. Configuring TLS Interception ============================ There are three ways to use Cilium in 1.17 and later: \* Using SDS, Secrets referenced in Network Policy can be located anywhere in the cluster, and are copied into a configured namespace | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.024418490007519722,
0.014302804134786129,
-0.011440346948802471,
-0.0351271815598011,
-0.05983072891831398,
-0.09166226536035538,
0.020785663276910782,
-0.025523511692881584,
0.113705113530159,
-0.021636415272951126,
0.035085007548332214,
-0.050834134221076965,
0.05358809232711792,
0.01... | 0.155434 |
method, SDS is the default for new Cilium installations as at Cilium 1.17. Configuring TLS Interception ============================ There are three ways to use Cilium in 1.17 and later: \* Using SDS, Secrets referenced in Network Policy can be located anywhere in the cluster, and are copied into a configured namespace (``cilium-secrets`` by default) by the Cilium Operator, synchronized from there into SDS, then referenced in NPDS using that name. This is the default for new clusters, and the recommended method of operation. \* Secrets can be located anywhere in the cluster, and the Cilium Agent can be granted read access to all Secrets in the cluster. In this case, Secrets are read directly from their original location by the Cilium Agent and sent inline in NPDS. This deployment method is included for backwards compatibility purposes and is \*\*not recommended\*\*, as it \*\*significantly expands\*\* the security scope of the agent. \* Secrets can be added directly to the ``cilium-secrets`` namespace, then referenced in that namespace from Network Policy. This is also included for backwards compatibility based on user feedback about how this feature was actually being used. It is the default for \*\*upgraded\*\* clusters that have not configured any settings and are using the ``upgradeCompatibility`` setting in Helm, set to ``1.16`` or below. There are three settings in Helm that affect TLS Interception: \* ``tls.secretsNamespace.name`` - default ``cilium-secrets``. Configures the secrets namespace that will be used for Policy secrets. Note that this is set to the same value as a similar setting for Ingress, Gateway API, and BGP configuration by default, but \*\*may\*\* be set to a different value. \* ``tls.readSecretsOnlyFromSecretsNamespace`` - default ``true``. This setting tells the Helm chart and Cilium whether the Cilium Agent should only read secrets from the configured Secrets namespace, or if the Cilium Agent should attempt to read Secrets directly from their location in the cluster. Previous versions of Cilium used the item ``tls.secretsBackend``, which could be set to ``local`` (meaning only read from the Secrets namespace) or ``k8s`` (meaning read from any namespace), but that field is now \*\*deprecated\*\*, as its naming had become detached from its function. Previous installations that set ``tls.secretsBackend`` to ``k8s`` should migrate to setting ``tls.readSecretsOnlyFromSecretsNamespace`` to ``false`` instead, although the setting will continue to work for Cilium 1.17. ``tls.secretsBackend`` will be removed in a future Cilium version. \* ``tls.secretSync.enabled`` - default ``true`` for new clusters. Configures secret synchronization and SDS use for Network Policy secrets. SDS use requires this to be set to ``true``, and must be disabled when this field is set to ``false``, so having an additional field for SDS config added no value. Configuring the three available modes for TLS Interception ---------------------------------------------------------- SDS Mode (recommended, default for new clusters): ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Set the following settings in your Helm Values: .. code-block:: YAML tls: readSecretsOnlyFromSecretsNamespace: true secretsNamespace: name: cilium-secrets # This setting is optional, as it is the default secretSync: enabled: true Read all Secrets in the Cluster mode (not recommended) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Set the following settings in your Helm Values: .. code-block:: YAML tls: readSecretsOnlyFromSecretsNamespace: false secretSync: enabled: false Read Secrets only from secrets namespace, no SDS (default for upgraded clusters) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Set the following settings in your Helm Values: .. code-block:: YAML tls: readSecretsOnlyFromSecretsNamespace: true secretsNamespace: name: cilium-secrets # This setting is optional, as it is the default secretSync: enabled: false .. Note:: If you are using this mode, then you will need to replace all references to ``kube-system`` in the validation instructions on this page with ``cilium-secrets`` (or whatever value you have set that namespace to). Once you've chosen an option and configured your Cilium installation | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.04562123492360115,
-0.005115709733217955,
-0.09107163548469543,
0.0013104036916047335,
0.013910939916968346,
-0.0694613829255104,
-0.02996111288666725,
0.000051772607548628,
0.03540710359811783,
0.005380045622587204,
0.051341891288757324,
-0.036754000931978226,
0.040462926030159,
-0.048... | 0.201552 |
default secretSync: enabled: false .. Note:: If you are using this mode, then you will need to replace all references to ``kube-system`` in the validation instructions on this page with ``cilium-secrets`` (or whatever value you have set that namespace to). Once you've chosen an option and configured your Cilium installation accordingly, proceed with verifying your install using the rest of these instructions. Deploy the Demo Application =========================== To demonstrate TLS-interception we will use the same ``mediabot`` application that we used for the DNS-aware policy example. This application will access the Star Wars API service using HTTPS, which would normally mean that network-layer mechanisms like Cilium would not be able to see the HTTP-layer details of the communication, since all application data is encrypted using TLS before that data is sent on the network. In this guide we will learn about: - Creating an internal Certificate Authority (CA) and associated certificates signed by that CA to enable TLS interception. - Using Cilium network policy to select the traffic to intercept using DNS-based policy rules. - Inspecting the details of the HTTP request using cilium monitor (accessing this visibility data via Hubble, and applying Cilium network policies to filter/modify the HTTP request is also possible, but is beyond the scope of this simple Getting Started Guide) First off, we will create a single pod ``mediabot`` application: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-sw-app.yaml $ kubectl wait pod/mediabot --for=condition=Ready $ kubectl get pods NAME READY STATUS RESTARTS AGE pod/mediabot 1/1 Running 0 14s Generating and Installing TLS Keys and Certificates =================================================== Now that we understand TLS and have configured Cilium to use TLS interception, we will walk through the concrete steps to generate the appropriate keys and certificates using the ``openssl`` utility. The following image describes the different files containing cryptographic data that are generated or copied, and what components in the system need access to those files: .. image:: images/cilium-tls-interception.png You can use openssl on your local system if it is already installed, but if not a simple shortcut is to use ``kubectl exec`` to execute ``/bin/bash`` within any of the cilium pods, and then run the resulting ``openssl`` commands. Use ``kubectl cp`` to copy the resulting files out of the cilium pod when it is time to use them to create Kubernetes secrets of copy them to the ``mediabot`` pod. Create an Internal Certificate Authority (CA) --------------------------------------------- Generate CA private key named 'myCA.key': .. code-block:: shell-session $ openssl genrsa -des3 -out myCA.key 2048 Enter any password, just remember it for some of the later steps. Generate CA certificate from the private key: .. code-block:: shell-session $ openssl req -x509 -new -nodes -key myCA.key -sha256 -days 1825 -out myCA.crt The values you enter for each prompt do not need to be any specific value, and do not need to be accurate. Create Private Key and Certificate Signing Request for a Given DNS Name ----------------------------------------------------------------------- Generate an internal private key and certificate signing with a common name that matches the DNS name of the destination service to be intercepted for inspection (in this example, use ``httpbin.org``). First create the private key: .. code-block:: shell-session $ openssl genrsa -out internal-httpbin.key 2048 Next, create a certificate signing request, specifying the DNS name of the destination service for the common name field when prompted. All other prompts can be filled with any value. .. code-block:: shell-session $ openssl req -new -key internal-httpbin.key -out internal-httpbin.csr The only field that must be a specific value is ensuring that ``Common Name`` is the exact DNS destination ``httpbin.org`` that will be provided to the client. This example workflow will | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.04596521705389023,
-0.02592882141470909,
-0.016307227313518524,
-0.113420270383358,
-0.03128630667924881,
-0.08400161564350128,
-0.0514250285923481,
-0.019732801243662834,
0.05912351235747337,
-0.00041646885802038014,
0.025810308754444122,
-0.039791472256183624,
0.02494390495121479,
0.0... | 0.096234 |
other prompts can be filled with any value. .. code-block:: shell-session $ openssl req -new -key internal-httpbin.key -out internal-httpbin.csr The only field that must be a specific value is ensuring that ``Common Name`` is the exact DNS destination ``httpbin.org`` that will be provided to the client. This example workflow will work for any DNS name as long as the toFQDNs rule in the policy YAML (below) is also updated to match the DNS name in the certificate. Use CA to Generate a Signed Certificate for the DNS Name -------------------------------------------------------- Use the internal CA private key to create a signed certificate for httpbin.org named ``internal-httpbin.crt``. .. code-block:: shell-session $ openssl x509 -req -days 360 -in internal-httpbin.csr -CA myCA.crt -CAkey myCA.key -CAcreateserial -out internal-httpbin.crt -sha256 Next we create a Kubernetes secret that includes both the private key and signed certificates for the destination service: .. code-block:: shell-session $ kubectl create secret tls httpbin-tls-data -n kube-system --cert=internal-httpbin.crt --key=internal-httpbin.key Add the Internal CA as a Trusted CA Inside the Client Pod --------------------------------------------------------- Once the CA certificate is inside the client pod, we still must make sure that the CA file is picked up by the TLS library used by your application. Most Linux applications automatically use a set of trusted CA certificates that are bundled along with the Linux distro. In this guide, we are using an Ubuntu container as the client, and so will update it with Ubuntu specific instructions. Other Linux distros will have different mechanisms. Also, individual applications may leverage their own certificate stores rather than use the OS certificate store. Java applications and the aws-cli are two common examples. Please refer to the application or application runtime documentation for more details. For Ubuntu, we first copy the additional CA certificate to the client pod filesystem .. code-block:: shell-session $ kubectl cp myCA.crt default/mediabot:/usr/local/share/ca-certificates/myCA.crt Then run the Ubuntu-specific utility that adds this certificate to the global set of trusted certificate authorities in /etc/ssl/certs/ca-certificates.crt . .. code-block:: shell-session $ kubectl exec mediabot -- update-ca-certificates This command will issue a WARNING, but this can be ignored. Provide Cilium with List of Trusted CAs --------------------------------------- Next, we will provide Cilium with the set of CAs that it should trust when originating the secondary TLS connections. This list should correspond to the standard set of global CAs that your organization trusts. A logical option for this is the standard CAs that are trusted by your operating system, since this is the set of CAs that were being used prior to introducing TLS inspection. To keep things simple, in this example we will simply copy this list out of the Ubuntu filesystem of the mediabot pod, though it is important to understand that this list of trusted CAs is not specific to a particular TLS client or server, and so this step need only be performed once regardless of how many TLS clients or servers are involved in TLS inspection. .. code-block:: shell-session $ kubectl cp default/mediabot:/etc/ssl/certs/ca-certificates.crt ca-certificates.crt We then will create a Kubernetes secret using this certificate bundle so that Cilium can read the certificate bundle and use it to validate outgoing TLS connections. .. code-block:: shell-session $ kubectl create secret generic tls-orig-data -n kube-system --from-file=ca.crt=./ca-certificates.crt Apply DNS and TLS-aware Egress Policy ===================================== Up to this point, we have created keys and certificates to enable TLS inspection, but we have not told Cilium which traffic we want to intercept and inspect. This is done using the same Cilium Network Policy constructs that are used for other Cilium Network Policies. The following Cilium network policy indicates that Cilium should perform HTTP-aware inspect of communication between the | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.04810989275574684,
0.07029170542955399,
-0.07011577486991882,
-0.056668974459171295,
-0.07951599359512329,
-0.04511597007513046,
0.004148625303059816,
-0.0454506017267704,
0.12319160252809525,
-0.021276412531733513,
-0.05208870396018028,
-0.15087378025054932,
0.06825755536556244,
0.0256... | -0.059267 |
TLS inspection, but we have not told Cilium which traffic we want to intercept and inspect. This is done using the same Cilium Network Policy constructs that are used for other Cilium Network Policies. The following Cilium network policy indicates that Cilium should perform HTTP-aware inspect of communication between the ``mediabot`` pod to ``httpbin.org``. .. literalinclude:: ../../examples/kubernetes-tls-inspection/l7-visibility-tls.yaml :language: yaml Let's take a closer look at the policy: \* The ``endpointSelector`` means that this policy will only apply to pods with labels ``class: mediabot, org:empire`` to have the egress access. \* The first egress section uses ``toFQDNs: matchName`` specification to allow TCP port 443 egress to ``httpbin.org``. \* The ``http`` section below the toFQDNs rule indicates that such connections should be parsed as HTTP, with a policy of ``{}`` which will allow all requests. \* The ``terminatingTLS`` and ``originatingTLS`` sections indicate that TLS interception should be used to terminate the initial TLS connection from mediabot and initiate a new out-bound TLS connection to ``httpbin.org``. \* The second egress section allows ``mediabot`` pods to access ``kube-dns`` service. Note that ``rules: dns`` instructs Cilium to inspect and allow DNS lookups matching specified patterns. In this case, inspect and allow all DNS queries. Note that with this policy the ``mediabot`` doesn't have access to any internal cluster service other than ``kube-dns`` and will have no access to any other external destinations either. Refer to :ref:`Network Policy` to learn more about policies for controlling access to internal cluster services. Let's apply the policy: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-tls-inspection/l7-visibility-tls.yaml Demonstrating TLS Inspection ============================ Recall that the policy we pushed will allow all HTTPS requests from ``mediabot`` to ``httpbin.org``, but will parse all data at the HTTP-layer, meaning that cilium monitor will report each HTTP request and response. To see this, open a new window and run the following command to identity the name of the cilium pod (e.g, cilium-97s78) that is running on the same Kubernetes worker node as the ``mediabot`` pod. Then start running cilium-dbg monitor in "L7 mode" to monitor for HTTP requests being reported by Cilium: .. code-block:: shell-session $ kubectl exec -it -n kube-system cilium-d5x8v -- cilium-dbg monitor -t l7 Next in the original window, from the ``mediabot`` pod we can access ``httpbin.org`` via HTTPS: .. code-block:: shell-session $ kubectl exec -it mediabot -- curl -sL 'https://httpbin.org/anything' ... ... $ kubectl exec -it mediabot -- curl -sL 'https://httpbin.org/headers' ... ... Looking back at the cilium-dbg monitor window, you will see each individual HTTP request and response. For example:: -> Request http from 2585 ([k8s:class=mediabot k8s:org=empire k8s:io.kubernetes.pod.namespace=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.cilium.k8s.policy.cluster=default]) to 0 ([reserved:world]), identity 24948->2, verdict Forwarded GET https://httpbin.org/anything => 0 -> Response http to 2585 ([k8s:io.kubernetes.pod.namespace=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.cilium.k8s.policy.cluster=default k8s:class=mediabot k8s:org=empire]) from 0 ([reserved:world]), identity 24948->2, verdict Forwarded GET https://httpbin.org/anything => 200 Refer to :ref:`l4\_policy` and :ref:`l7\_policy` to learn more about Cilium L4 and L7 network policies. Clean-up ======== .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-dns/dns-sw-app.yaml $ kubectl delete cnp l7-visibility-tls $ kubectl delete secret -n kube-system tls-orig-data $ kubectl delete secret -n kube-system httpbin-tls-data | https://github.com/cilium/cilium/blob/main//Documentation/security/tls-visibility.rst | main | cilium | [
-0.016343465074896812,
0.04277607053518295,
-0.010500425472855568,
-0.06644725799560547,
-0.005482611712068319,
-0.04933939874172211,
0.03139445185661316,
-0.057522423565387726,
0.08803465962409973,
0.030067963525652885,
0.005258924327790737,
-0.06604144722223282,
0.012853620573878288,
-0.... | 0.138571 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_aws\_metadata\_with\_policy: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Locking Down External Access Using AWS Metadata \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction to using Cilium to enforce policies based on AWS metadata. It provides a detailed walk-through of running a single-node Cilium environment on your machine. It is designed to take 15-30 minutes for users with some experience running Kubernetes. Setup Cilium ============ This guide will work with any approach to installing Cilium, including minikube, as long as the cilium-operator pod in the deployment can reach the AWS API server However, since the most common use of this mechanism is for Kubernetes clusters running in AWS, we recommend trying it out along with the guide: :ref:`k8s\_install\_quick` . Create AWS secrets ================== Before installing Cilium, a new Kubernetes Secret with the AWS Tokens needs to be added to your Kubernetes cluster. This Secret will allow Cilium to gather information from the AWS API which is needed to implement ToGroups policies. AWS Access keys and IAM role ------------------------------ To create a new access token the `following guide can be used `\_. These keys need to have certain permissions set: .. code-block:: javascript { "Version": "2012-10-17", "Statement": [ { "Effect": "Allow", "Action": "ec2:Describe\*", "Resource": "\*" } ] } As soon as you have the access tokens, the following secret needs to be added, with each empty string replaced by the associated value as a base64-encoded string: .. code-block:: yaml :name: cilium-secret.yaml apiVersion: v1 kind: Secret metadata: name: cilium-aws namespace: kube-system type: Opaque data: AWS\_ACCESS\_KEY\_ID: "" AWS\_SECRET\_ACCESS\_KEY: "" AWS\_DEFAULT\_REGION: "" The base64 command line utility can be used to generate each value, for example: .. code-block:: shell-session $ echo -n "eu-west-1" | base64 ZXUtd2VzdC0x This secret stores the AWS credentials, which will be used to connect the AWS API. .. code-block:: shell-session $ kubectl create -f cilium-secret.yaml To validate that the credentials are correct, the following pod can be created for debugging purposes: .. code-block:: yaml apiVersion: v1 kind: Pod metadata: name: testing-aws-pod namespace: kube-system spec: containers: - name: aws-cli image: mesosphere/aws-cli command: ['sh', '-c', 'sleep 3600'] env: - name: AWS\_ACCESS\_KEY\_ID valueFrom: secretKeyRef: name: cilium-aws key: AWS\_ACCESS\_KEY\_ID optional: true - name: AWS\_SECRET\_ACCESS\_KEY valueFrom: secretKeyRef: name: cilium-aws key: AWS\_SECRET\_ACCESS\_KEY optional: true - name: AWS\_DEFAULT\_REGION valueFrom: secretKeyRef: name: cilium-aws key: AWS\_DEFAULT\_REGION optional: true To list all of the available AWS instances, the following command can be used: .. code-block:: shell-session $ kubectl -n kube-system exec -ti testing-aws-pod -- aws ec2 describe-instances Once the secret has been created and validated, the cilium-operator pod must be restarted in order to pick up the credentials in the secret. To do this, identify and delete the existing cilium-operator pod, which will be recreated automatically: .. code-block:: shell-session $ kubectl get pods -l name=cilium-operator -n kube-system NAME READY STATUS RESTARTS AGE cilium-operator-7c9d69f7c-97vqx 1/1 Running 0 36h $ kubectl delete pod cilium-operator-7c9d69f7c-97vqx It is important for this demo that ``coredns`` is working correctly. To know the status of ``coredns`` you can run the following command: .. code-block:: shell-session $ kubectl get deployment -l k8s-app=kube-dns -n kube-system NAME DESIRED CURRENT UP-TO-DATE AVAILABLE AGE coredns 2 2 2 2 13h Where at least one pod should be available. Configure AWS Security Groups ============================= Cilium's AWS Metadata filtering capability enables explicit whitelisting of communication between a subset of pods (identified by Kubernetes labels) with a set of destination EC2 ENIs (identified by membership in an AWS security group). In this example, the destination EC2 elastic network interfaces are attached to EC2 instances that are members | https://github.com/cilium/cilium/blob/main//Documentation/security/aws.rst | main | cilium | [
-0.008256109431385994,
0.05226975679397583,
-0.1182752326130867,
-0.04082825779914856,
0.08579869568347931,
-0.06268858164548874,
-0.054885245859622955,
-0.013232422061264515,
0.07888798415660858,
0.05433298647403717,
0.03205202519893646,
-0.07058864831924438,
0.042433831840753555,
-0.0805... | 0.104553 |
AWS Metadata filtering capability enables explicit whitelisting of communication between a subset of pods (identified by Kubernetes labels) with a set of destination EC2 ENIs (identified by membership in an AWS security group). In this example, the destination EC2 elastic network interfaces are attached to EC2 instances that are members of a single AWS security group ('sg-0f2146100a88d03c3'). Pods with label ``class=xwing`` should only be able to make connections outside the cluster to the destination network interfaces in that security group. To enable this, the VMs acting as Kubernetes worker nodes must be able to send traffic to the destination VMs that are being accessed by pods. One approach for achieving this is to put all Kubernetes worker VMs in a single 'k8s-worker' security group, and then ensure that any security group that is referenced in a Cilium toGroups policy has an allow all ingress rule (all ports) for connections from the 'k8s-worker' security group. Cilium filtering will then ensure that the only pods allowed by policy can reach the destination VMs. Create a sample policy ====================== Deploy a demo application: ---------------------------- In this case we're going to use a demo application that is used in other guides. These manifests will create three microservices applications: \*deathstar\*, \*tiefighter\*, and \*xwing\*. In this case, we are only going to use our \*xwing\* microservice to secure communications to existing AWS instances. .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/minikube/http-sw-app.yaml service "deathstar" created deployment "deathstar" created deployment "tiefighter" created deployment "xwing" created Kubernetes will deploy the pods and service in the background. Running ``kubectl get pods,svc`` will inform you about the progress of the operation. Each pod will go through several states until it reaches ``Running`` at which point the pod is ready. .. code-block:: shell-session $ kubectl get pods,svc NAME READY STATUS RESTARTS AGE po/deathstar-76995f4687-2mxb2 1/1 Running 0 1m po/deathstar-76995f4687-xbgnl 1/1 Running 0 1m po/tiefighter 1/1 Running 0 1m po/xwing 1/1 Running 0 1m NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE svc/deathstar ClusterIP 10.109.254.198 80/TCP 3h svc/kubernetes ClusterIP 10.96.0.1 443/TCP 3h Policy Language: ----------------- \*\*ToGroups\*\* rules can be used to define policy in relation to cloud providers, like AWS. .. code-block:: yaml --- kind: CiliumNetworkPolicy apiVersion: cilium.io/v2 metadata: name: to-groups-sample namespace: default spec: endpointSelector: matchLabels: org: alliance class: xwing egress: - toPorts: - ports: - port: '80' protocol: TCP toGroups: - aws: securityGroupsIds: - 'sg-0f2146100a88d03c3' This policy allows traffic from pod \*xwing\* to any AWS elastic network interface in the security group with ID ``sg-0f2146100a88d03c3``. Validate that derived policy is in place ---------------------------------------- Every time that a new policy with ToGroups rules is added, an equivalent policy (also called "derivative policy"), will be created. This policy will contain the set of CIDRs that correspond to the specification in ToGroups, e.g., the IPs of all network interfaces that are part of a specified security group. The list of IPs is updated periodically. .. code-block:: shell-session $ kubectl get cnp NAME AGE to-groups-sample 11s to-groups-sample-togroups-044ba7d1-f491-11e8-ad2e-080027d2d952 10s Eventually, the derivative policy will contain IPs in the ToCIDR section: .. code-block:: shell-session $ kubectl get cnp to-groups-sample-togroups-044ba7d1-f491-11e8-ad2e-080027d2d952 .. code-block:: yaml apiVersion: cilium.io/v2 kind: CiliumNetworkPolicy metadata: creationTimestamp: 2018-11-30T11:13:52Z generation: 1 labels: io.cilium.network.policy.kind: derivative io.cilium.network.policy.parent.uuid: 044ba7d1-f491-11e8-ad2e-080027d2d952 name: to-groups-sample-togroups-044ba7d1-f491-11e8-ad2e-080027d2d952 namespace: default ownerReferences: - apiVersion: cilium.io/v2 blockOwnerDeletion: true kind: CiliumNetworkPolicy name: to-groups-sample uid: 044ba7d1-f491-11e8-ad2e-080027d2d952 resourceVersion: "34853" selfLink: /apis/cilium.io/v2/namespaces/default/ciliumnetworkpolicies/to-groups-sample-togroups-044ba7d1-f491-11e8-ad2e-080027d2d952 uid: 04b289ba-f491-11e8-ad2e-080027d2d952 specs: - egress: - toCIDRSet: - cidr: 34.254.113.42/32 - cidr: 172.31.44.160/32 toPorts: - ports: - port: "80" protocol: TCP endpointSelector: matchLabels: any:class: xwing any:org: alliance k8s:io.kubernetes.pod.namespace: default labels: - key: io.cilium.k8s.policy.name source: k8s value: to-groups-sample - key: io.cilium.k8s.policy.uid source: k8s value: 044ba7d1-f491-11e8-ad2e-080027d2d952 - key: io.cilium.k8s.policy.namespace source: k8s value: default - key: io.cilium.k8s.policy.derived-from source: | https://github.com/cilium/cilium/blob/main//Documentation/security/aws.rst | main | cilium | [
0.05716030299663544,
0.07209469377994537,
-0.0226434338837862,
-0.034280925989151,
0.054106369614601135,
0.0023314058780670166,
0.09946063160896301,
-0.07956086844205856,
0.013832641765475273,
0.023253368213772774,
-0.03454321250319481,
-0.09714323282241821,
0.08371827751398087,
-0.0252804... | 0.134411 |
- toCIDRSet: - cidr: 34.254.113.42/32 - cidr: 172.31.44.160/32 toPorts: - ports: - port: "80" protocol: TCP endpointSelector: matchLabels: any:class: xwing any:org: alliance k8s:io.kubernetes.pod.namespace: default labels: - key: io.cilium.k8s.policy.name source: k8s value: to-groups-sample - key: io.cilium.k8s.policy.uid source: k8s value: 044ba7d1-f491-11e8-ad2e-080027d2d952 - key: io.cilium.k8s.policy.namespace source: k8s value: default - key: io.cilium.k8s.policy.derived-from source: k8s value: CiliumNetworkPolicy status: nodes: k8s1: enforcing: true lastUpdated: 2018-11-30T11:28:03.907678888Z localPolicyRevision: 28 ok: true The derivative rule should contain the following information: - \*metadata.OwnerReferences\*: that contains the information about the ToGroups policy. - \*specs.Egress.ToCIDRSet\*: the list of private and public IPs of the instances that correspond to the spec of the parent policy. - \*status\*: whether or not the policy is enforced yet, and when the policy was last updated. The endpoint status for the \*xwing\* should have policy enforcement enabled only for egress connectivity: .. code-block:: shell-session $ kubectl exec -q -it -n kube-system cilium-85vtg -- cilium-dbg endpoint get 23453 -o jsonpath='{$[0].status.policy.realized.policy-enabled}' egress In this example, \*xwing\* pod can only connect to ``34.254.113.42/32`` and ``172.31.44.160/32`` and connectivity to other IP will be denied. | https://github.com/cilium/cilium/blob/main//Documentation/security/aws.rst | main | cilium | [
-0.01903320848941803,
0.047135964035987854,
-0.0451110303401947,
-0.08304181694984436,
0.025574084371328354,
-0.047036342322826385,
0.012340983375906944,
-0.07091034203767776,
0.06994868814945221,
0.06509706377983093,
0.05381176993250847,
-0.09975293278694153,
-0.00756627693772316,
-0.0304... | 0.172477 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_policy\_verdicts: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Creating Policies from Verdicts \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Policy Audit Mode configures Cilium to allow all traffic while logging all connections that would otherwise be dropped by network policies. Policy Audit Mode may be configured for the entire daemon using ``--policy-audit-mode=true`` or for individual Cilium Endpoints. When Policy Audit Mode is enabled, no network policy is enforced so this setting is \*\*not recommended for production deployment\*\*. Policy Audit Mode supports auditing network policies implemented at networks layers 3 and 4. This guide walks through the process of creating policies using Policy Audit Mode. .. include:: gsg\_requirements.rst .. include:: gsg\_sw\_demo.rst Scale down the deathstar Deployment =================================== In this guide we're going to scale down the deathstar Deployment in order to simplify the next steps: .. code-block:: shell-session $ kubectl scale --replicas=1 deployment deathstar deployment.apps/deathstar scaled Enable Policy Audit Mode (Entire Daemon) ======================================== To observe policy audit messages for all endpoints managed by this Daemonset, modify the Cilium ConfigMap and restart all daemons: .. tabs:: .. group-tab:: Configure via kubectl .. code-block:: shell-session $ kubectl patch -n $CILIUM\_NAMESPACE configmap cilium-config --type merge --patch '{"data":{"policy-audit-mode":"true"}}' configmap/cilium-config patched $ kubectl -n $CILIUM\_NAMESPACE rollout restart ds/cilium daemonset.apps/cilium restarted $ kubectl -n $CILIUM\_NAMESPACE rollout status ds/cilium Waiting for daemon set "cilium" rollout to finish: 0 of 1 updated pods are available... daemon set "cilium" successfully rolled out .. group-tab:: Helm Upgrade If you installed Cilium via ``helm install``, then you can use ``helm upgrade`` to enable Policy Audit Mode: .. cilium-helm-upgrade:: :namespace: $CILIUM\_NAMESPACE :extra-args: --reuse-values :set: policyAuditMode=true Enable Policy Audit Mode (Specific Endpoint) ============================================ Cilium can enable Policy Audit Mode for a specific endpoint. This may be helpful when enabling Policy Audit Mode for the entire daemon is too broad. Enabling per endpoint will ensure other endpoints managed by the same daemon are not impacted. This approach is meant to be temporary. \*\*Restarting Cilium pod will reset the Policy Audit Mode to match the daemon's configuration.\*\* Policy Audit Mode is enabled for a given endpoint by modifying the endpoint configuration via the ``cilium-dbg`` tool on the endpoint's Kubernetes node. The steps include: #. Determine the endpoint id on which Policy Audit Mode will be enabled. #. Identify the Cilium pod running on the same Kubernetes node corresponding to the endpoint. #. Using the Cilium pod above, modify the endpoint configuration by setting ``PolicyAuditMode=Enabled``. The following shell commands perform these steps: .. code-block:: shell-session $ PODNAME=$(kubectl get pods -l app.kubernetes.io/name=deathstar -o jsonpath='{.items[\*].metadata.name}') $ NODENAME=$(kubectl get pod -o jsonpath="{.items[?(@.metadata.name=='$PODNAME')].spec.nodeName}") $ ENDPOINT=$(kubectl get cep -o jsonpath="{.items[?(@.metadata.name=='$PODNAME')].status.id}") $ CILIUM\_POD=$(kubectl -n "$CILIUM\_NAMESPACE" get pod --all-namespaces --field-selector spec.nodeName="$NODENAME" -lk8s-app=cilium -o jsonpath='{.items[\*].metadata.name}') $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ cilium-dbg endpoint config "$ENDPOINT" PolicyAuditMode=Enabled Endpoint 232 configuration updated successfully We can check that Policy Audit Mode is enabled for this endpoint with .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ cilium-dbg endpoint get "$ENDPOINT" -o jsonpath='{[\*].spec.options.PolicyAuditMode}' Enabled .. \_observe\_policy\_verdicts: Observe policy verdicts ======================= In this example, we are tasked with applying security policy for the deathstar. First, from the Cilium pod we need to monitor the notifications for policy verdicts using the Hubble CLI. We'll be monitoring for inbound traffic towards the deathstar to identify it and determine whether to extend the network policy to allow that traffic. Apply a default-deny policy: .. literalinclude:: ../../examples/minikube/sw\_deny\_policy.yaml :language: yaml CiliumNetworkPolicies match on pod labels using an ``endpointSelector`` to identify the sources and destinations to which the policy applies. The above policy denies | https://github.com/cilium/cilium/blob/main//Documentation/security/policy-creation.rst | main | cilium | [
0.00045216272701509297,
0.06071023270487785,
-0.07444391399621964,
-0.07289936393499374,
0.06364437937736511,
-0.025028010830283165,
0.049834687262773514,
-0.009594994597136974,
0.02346189320087433,
0.03544732555747032,
0.06255553662776947,
-0.023194575682282448,
0.03821069747209549,
-0.01... | 0.166933 |
towards the deathstar to identify it and determine whether to extend the network policy to allow that traffic. Apply a default-deny policy: .. literalinclude:: ../../examples/minikube/sw\_deny\_policy.yaml :language: yaml CiliumNetworkPolicies match on pod labels using an ``endpointSelector`` to identify the sources and destinations to which the policy applies. The above policy denies traffic sent to any pods with label (``org=empire``). Due to the Policy Audit Mode enabled above (either for the entire daemon, or for just the ``deathstar`` endpoint), the traffic will not actually be denied but will instead trigger policy verdict notifications. To apply this policy, run: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/minikube/sw\_deny\_policy.yaml ciliumnetworkpolicy.cilium.io/empire-default-deny created With the above policy, we will enable a default-deny posture on ingress to pods with the label ``org=empire`` and enable the policy verdict notifications for those pods. The same principle applies on egress as well. Now let's send some traffic from the tiefighter to the deathstar: .. code-block:: shell-session $ kubectl exec tiefighter -- curl -s -XPOST deathstar.default.svc.cluster.local/v1/request-landing Ship landed We can check the policy verdict from the Cilium Pod: .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ hubble observe flows -t policy-verdict --last 1 Feb 7 12:53:39.168: default/tiefighter:54134 (ID:31028) -> default/deathstar-6fb5694d48-5hmds:80 (ID:16530) policy-verdict:none AUDITED (TCP Flags: SYN) In the above example, we can see that the Pod ``deathstar-6fb5694d48-5hmds`` has received traffic from the ``tiefighter`` Pod which doesn't match the policy (``policy-verdict:none AUDITED``). .. \_create\_network\_policy: Create the Network Policy ========================= We can get more information about the flow with .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ hubble observe flows -t policy-verdict -o json --last 1 Given the above information, we now know the labels of the source and destination Pods, the traffic direction, and the destination port. In this case, we can see clearly that the source (i.e. the tiefighter Pod) is an empire aircraft (as it has the ``org=empire`` label) so once we've determined that we expect this traffic to arrive at the deathstar, we can form a policy to match the traffic: .. literalinclude:: ../../examples/minikube/sw\_l3\_l4\_policy.yaml :language: yaml To apply this L3/L4 policy, run: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/minikube/sw\_l3\_l4\_policy.yaml ciliumnetworkpolicy.cilium.io/rule1 created Now if we run the landing requests again, .. code-block:: shell-session $ kubectl exec tiefighter -- curl -s -XPOST deathstar.default.svc.cluster.local/v1/request-landing Ship landed we can then observe that the traffic which was previously audited to be dropped by the policy is reported as allowed: .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ hubble observe flows -t policy-verdict --last 1 ... Feb 7 13:06:45.130: default/tiefighter:59824 (ID:31028) -> default/deathstar-6fb5694d48-5hmds:80 (ID:16530) policy-verdict:L3-L4 ALLOWED (TCP Flags: SYN) Now the policy verdict states that the traffic would be allowed: ``policy-verdict:L3-L4 ALLOWED``. Success! Disable Policy Audit Mode (Entire Daemon) ========================================= These steps should be repeated for each connection in the cluster to ensure that the network policy allows all of the expected traffic. The final step after deploying the policy is to disable Policy Audit Mode again: .. tabs:: .. group-tab:: Configure via kubectl .. code-block:: shell-session $ kubectl patch -n $CILIUM\_NAMESPACE configmap cilium-config --type merge --patch '{"data":{"policy-audit-mode":"false"}}' configmap/cilium-config patched $ kubectl -n $CILIUM\_NAMESPACE rollout restart ds/cilium daemonset.apps/cilium restarted $ kubectl -n $CILIUM\_NAMESPACE rollout status ds/cilium Waiting for daemon set "cilium" rollout to finish: 0 of 1 updated pods are available... daemon set "cilium" successfully rolled out .. group-tab:: Helm Upgrade .. cilium-helm-upgrade:: :namespace: $CILIUM\_NAMESPACE :extra-args: --reuse-values :set: policyAuditMode=false Disable Policy Audit Mode (Specific Endpoint) ============================================= These steps are nearly identical to enabling Policy Audit Mode. .. code-block:: shell-session $ PODNAME=$(kubectl get pods -l app.kubernetes.io/name=deathstar -o jsonpath='{.items[\*].metadata.name}') $ NODENAME=$(kubectl get | https://github.com/cilium/cilium/blob/main//Documentation/security/policy-creation.rst | main | cilium | [
0.0481053851544857,
0.041781194508075714,
0.0004285838222131133,
-0.08157380670309067,
0.06999131292104721,
0.014447616413235664,
0.03671020269393921,
-0.020335664972662926,
0.029504749923944473,
0.09606216102838516,
0.021464340388774872,
-0.03869912400841713,
0.01937561295926571,
-0.02961... | 0.096253 |
available... daemon set "cilium" successfully rolled out .. group-tab:: Helm Upgrade .. cilium-helm-upgrade:: :namespace: $CILIUM\_NAMESPACE :extra-args: --reuse-values :set: policyAuditMode=false Disable Policy Audit Mode (Specific Endpoint) ============================================= These steps are nearly identical to enabling Policy Audit Mode. .. code-block:: shell-session $ PODNAME=$(kubectl get pods -l app.kubernetes.io/name=deathstar -o jsonpath='{.items[\*].metadata.name}') $ NODENAME=$(kubectl get pod -o jsonpath="{.items[?(@.metadata.name=='$PODNAME')].spec.nodeName}") $ ENDPOINT=$(kubectl get cep -o jsonpath="{.items[?(@.metadata.name=='$PODNAME')].status.id}") $ CILIUM\_POD=$(kubectl -n "$CILIUM\_NAMESPACE" get pod --all-namespaces --field-selector spec.nodeName="$NODENAME" -lk8s-app=cilium -o jsonpath='{.items[\*].metadata.name}') $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ cilium-dbg endpoint config "$ENDPOINT" PolicyAuditMode=Disabled Endpoint 232 configuration updated successfully Alternatively, \*\*restarting the Cilium pod\*\* will set the endpoint Policy Audit Mode to the daemon set configuration. Verify Policy Audit Mode is Disabled ==================================== .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ cilium-dbg endpoint get "$ENDPOINT" -o jsonpath='{[\*].spec.options.PolicyAuditMode}' Disabled Now if we run the landing requests again, only the \*tiefighter\* pods with the label ``org=empire`` should succeed: .. code-block:: shell-session $ kubectl exec tiefighter -- curl -s -XPOST deathstar.default.svc.cluster.local/v1/request-landing Ship landed And we can observe that the traffic was allowed by the policy: .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ hubble observe flows -t policy-verdict --from-pod tiefighter --last 1 Feb 7 13:34:26.112: default/tiefighter:37314 (ID:31028) -> default/deathstar-6fb5694d48-5hmds:80 (ID:16530) policy-verdict:L3-L4 ALLOWED (TCP Flags: SYN) This works as expected. Now the same request from an \*xwing\* Pod should fail: .. code-block:: shell-session $ kubectl exec xwing -- curl --connect-timeout 3 -s -XPOST deathstar.default.svc.cluster.local/v1/request-landing command terminated with exit code 28 This curl request should timeout after three seconds, we can observe the policy verdict with: .. code-block:: shell-session $ kubectl -n "$CILIUM\_NAMESPACE" exec "$CILIUM\_POD" -c cilium-agent -- \ hubble observe flows -t policy-verdict --from-pod xwing --last 1 Feb 7 13:43:46.791: default/xwing:54842 (ID:22654) <> default/deathstar-6fb5694d48-5hmds:80 (ID:16530) policy-verdict:none DENIED (TCP Flags: SYN) We hope you enjoyed the tutorial. Feel free to play more with the setup, follow the `gs\_http` guide, and reach out to us on `Cilium Slack`\_ with any questions! Clean-up ======== .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/minikube/http-sw-app.yaml $ kubectl delete cnp empire-default-deny rule1 | https://github.com/cilium/cilium/blob/main//Documentation/security/policy-creation.rst | main | cilium | [
0.0636066198348999,
0.08727016299962997,
-0.011605843901634216,
-0.02598538063466549,
0.05914454907178879,
-0.03029133751988411,
-0.0042005483992397785,
-0.04049814119935036,
0.07169239968061447,
0.06996744871139526,
0.06545242667198181,
-0.08801817893981934,
-0.01694900169968605,
-0.03543... | 0.131522 |
Deploy the Demo Application =========================== When we have Cilium deployed and ``kube-dns`` operating correctly we can deploy our demo application. In our Star Wars-inspired example, there are three microservices applications: \*deathstar\*, \*tiefighter\*, and \*xwing\*. The \*deathstar\* runs an HTTP webservice on port 80, which is exposed as a `Kubernetes Service `\_ to load-balance requests to \*deathstar\* across two pod replicas. The \*deathstar\* service provides landing services to the empire's spaceships so that they can request a landing port. The \*tiefighter\* pod represents a landing-request client service on a typical empire ship and \*xwing\* represents a similar service on an alliance ship. They exist so that we can test different security policies for access control to \*deathstar\* landing services. \*\*Application Topology for Cilium and Kubernetes\*\* .. image:: /gettingstarted/images/cilium\_http\_gsg.png :scale: 30 % The file ``http-sw-app.yaml`` contains a `Kubernetes Deployment `\_ for each of the three services. Each deployment is identified using the Kubernetes labels (``org=empire, class=deathstar``), (``org=empire, class=tiefighter``), and (``org=alliance, class=xwing``). It also includes a deathstar-service, which load-balances traffic to all pods with label (``org=empire, class=deathstar``). .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/minikube/http-sw-app.yaml service/deathstar created deployment.apps/deathstar created pod/tiefighter created pod/xwing created Kubernetes will deploy the pods and service in the background. Running ``kubectl get pods,svc`` will inform you about the progress of the operation. Each pod will go through several states until it reaches ``Running`` at which point the pod is ready. .. code-block:: shell-session $ kubectl get pods,svc NAME READY STATUS RESTARTS AGE pod/deathstar-6fb5694d48-5hmds 1/1 Running 0 107s pod/deathstar-6fb5694d48-fhf65 1/1 Running 0 107s pod/tiefighter 1/1 Running 0 107s pod/xwing 1/1 Running 0 107s NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE service/deathstar ClusterIP 10.96.110.8 80/TCP 107s service/kubernetes ClusterIP 10.96.0.1 443/TCP 3m53s Each pod will be represented in Cilium as an :ref:`endpoint` in the local cilium agent. We can invoke the ``cilium-dbg`` tool inside the Cilium pod to list them (in a single-node installation ``kubectl -n kube-system exec ds/cilium -- cilium-dbg endpoint list`` lists them all, but in a multi-node installation, only the ones running on the same node will be listed): .. code-block:: shell-session $ kubectl -n kube-system get pods -l k8s-app=cilium NAME READY STATUS RESTARTS AGE cilium-5ngzd 1/1 Running 0 3m19s $ kubectl -n kube-system exec cilium-5ngzd -- cilium-dbg endpoint list ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 232 Disabled Disabled 16530 k8s:class=deathstar 10.0.0.147 ready k8s:io.cilium.k8s.policy.cluster=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.kubernetes.pod.namespace=default k8s:org=empire 726 Disabled Disabled 1 reserved:host ready 883 Disabled Disabled 4 reserved:health 10.0.0.244 ready 1634 Disabled Disabled 51373 k8s:io.cilium.k8s.policy.cluster=default 10.0.0.118 ready k8s:io.cilium.k8s.policy.serviceaccount=coredns k8s:io.kubernetes.pod.namespace=kube-system k8s:k8s-app=kube-dns 1673 Disabled Disabled 31028 k8s:class=tiefighter 10.0.0.112 ready k8s:io.cilium.k8s.policy.cluster=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.kubernetes.pod.namespace=default k8s:org=empire 2811 Disabled Disabled 51373 k8s:io.cilium.k8s.policy.cluster=default 10.0.0.47 ready k8s:io.cilium.k8s.policy.serviceaccount=coredns k8s:io.kubernetes.pod.namespace=kube-system k8s:k8s-app=kube-dns 2843 Disabled Disabled 16530 k8s:class=deathstar 10.0.0.89 ready k8s:io.cilium.k8s.policy.cluster=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.kubernetes.pod.namespace=default k8s:org=empire 3184 Disabled Disabled 22654 k8s:class=xwing 10.0.0.30 ready k8s:io.cilium.k8s.policy.cluster=default k8s:io.cilium.k8s.policy.serviceaccount=default k8s:io.kubernetes.pod.namespace=default k8s:org=alliance Both ingress and egress policy enforcement is still disabled on all of these pods because no network policy has been imported yet which select any of the pods. | https://github.com/cilium/cilium/blob/main//Documentation/security/gsg_sw_demo.rst | main | cilium | [
-0.046073149889707565,
-0.034058935940265656,
-0.015253156423568726,
-0.13623525202274323,
-0.08944608271121979,
-0.08063679188489914,
-0.01183006726205349,
-0.025076180696487427,
0.046299077570438385,
0.028513731434941292,
0.021988343447446823,
-0.04132680594921112,
0.013478162698447704,
... | 0.214831 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_security\_restrict\_pod\_access: ======================================== Restricting privileged Cilium pod access ======================================== This page shows you how to restrict privileged access to Cilium pods by limiting access from the Kubernetes API, specifically from `kubernetes exec pod`\_. .. include:: gsg\_requirements.rst Background ---------- The Cilium agent needs some specific Linux capabilities to perform essential system and network operations. Cilium relies on Kubernetes and containers to set up the environment and mount the corresponding volumes. Cilium doesn't perform any extra operations that could result in an unsafe volume mount. Cilium needs kernel interfaces to properly configure the environment. Some kernel interfaces are part of the ``/proc`` filesystem, which includes host and machine configurations that can't be virtualized or namespaced. If ``pod exec`` operations aren't restricted, then remote `exec into pods`\_ and containers defeats Linux namespace restrictions. The Linux kernel restricts joining other namespaces by default. To enter the Cilium container, the ``CAP\_SYS\_ADMIN`` capability is required in both the current user namespace and in the Cilium user namespace (the initial namespace). If both namespaces have the ``CAP\_SYS\_ADMIN`` capability, then this is already a privileged access. To prevent privileged access to Cilium pods, restrict access to the Kubernetes API and arbitrary ``pod exec`` operations. Restrict authorization for ``kubernetes exec pod`` -------------------------------------------------- To restrict access to Cilium pods through ``kubernetes exec pod``: 1. Configure `RBAC authorization`\_ in Kubernetes. 2. Limit `access to the proxy subresource of Nodes`\_. References ---------- For more information about namespace security, visit: - https://man7.org/linux/man-pages/man7/user\_namespaces.7.html - https://man7.org/linux/man-pages/man1/nsenter.1.html - https://man7.org/linux/man-pages/man2/setns.2.html .. \_kubernetes exec pod: https://kubernetes.io/docs/tasks/debug/debug-application/get-shell-running-container/ .. \_exec into pods: https://kubernetes.io/docs/tasks/debug/debug-application/get-shell-running-container/ .. \_RBAC authorization: https://kubernetes.io/docs/reference/access-authn-authz/rbac/ .. \_access to the proxy subresource of Nodes: https://kubernetes.io/docs/concepts/security/rbac-good-practices/#access-to-proxy-subresource-of-nodes | https://github.com/cilium/cilium/blob/main//Documentation/security/restrict-pod-access.rst | main | cilium | [
0.034796737134456635,
0.041864510625600815,
-0.04569975286722183,
-0.06588619947433472,
0.029244674369692802,
-0.05710740387439728,
-0.01827639900147915,
0.04097921773791313,
0.03363471105694771,
-0.024292796850204468,
0.03637304902076721,
-0.0952017605304718,
0.037128206342458725,
-0.0046... | 0.165981 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io .. \_gs\_kafka: \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* Securing a Kafka Cluster \*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\*\* This document serves as an introduction to using Cilium to enforce Kafka-aware security policies. It is a detailed walk-through of getting a single-node Cilium environment running on your machine. It is designed to take 15-30 minutes. .. include:: gsg\_requirements.rst Deploy the Demo Application =========================== Now that we have Cilium deployed and ``kube-dns`` operating correctly we can deploy our demo Kafka application. Since our first demo of Cilium + HTTP-aware security policies was Star Wars-themed we decided to do the same for Kafka. While the `HTTP-aware Cilium Star Wars demo `\_ showed how the Galactic Empire used HTTP-aware security policies to protect the Death Star from the Rebel Alliance, this Kafka demo shows how the lack of Kafka-aware security policies allowed the Rebels to steal the Death Star plans in the first place. Kafka is a powerful platform for passing datastreams between different components of an application. A cluster of "Kafka brokers" connect nodes that "produce" data into a data stream, or "consume" data from a datastream. Kafka refers to each datastream as a "topic". Because scalable and highly-available Kafka clusters are non-trivial to run, the same cluster of Kafka brokers often handles many different topics at once (read this `Introduction to Kafka `\_ for more background). In our simple example, the Empire uses a Kafka cluster to handle two different topics: - \*empire-announce\* : Used to broadcast announcements to sites spread across the galaxy - \*deathstar-plans\* : Used by a small group of sites coordinating on building the ultimate battlestation. To keep the setup small, we will just launch a small number of pods to represent this setup: - \*kafka-broker\* : A single pod running Kafka and Zookeeper representing the Kafka cluster (label app=kafka). - \*empire-hq\* : A pod representing the Empire's Headquarters, which is the only pod that should produce messages to \*empire-announce\* or \*deathstar-plans\* (label app=empire-hq). - \*empire-backup\* : A secure backup facility located in `Scarif `\_ , which is allowed to "consume" from the secret \*deathstar-plans\* topic (label app=empire-backup). - \*empire-outpost-8888\* : A random outpost in the empire. It needs to "consume" messages from the \*empire-announce\* topic (label app=empire-outpost). - \*empire-outpost-9999\* : Another random outpost in the empire that "consumes" messages from the \*empire-announce\* topic (label app=empire-outpost). All pods other than \*kafka-broker\* are Kafka clients, which need access to the \*kafka-broker\* container on TCP port 9092 in order to send Kafka protocol messages. .. image:: images/cilium\_kafka\_gsg\_topology.png The file ``kafka-sw-app.yaml`` contains a Kubernetes Deployment for each of the pods described above, as well as a Kubernetes Service for both Kafka and Zookeeper. .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-kafka/kafka-sw-app.yaml deployment "kafka-broker" created deployment "zookeeper" created service "zook" created service "kafka-service" created deployment "empire-hq" created deployment "empire-outpost-8888" created deployment "empire-outpost-9999" created deployment "empire-backup" created Kubernetes will deploy the pods and service in the background. Running ``kubectl get svc,pods`` will inform you about the progress of the operation. Each pod will go through several states until it reaches ``Running`` at which point the setup is ready. .. code-block:: shell-session $ kubectl get svc,pods NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kafka-service ClusterIP None 9092/TCP 2m kubernetes ClusterIP 10.96.0.1 443/TCP 10m zook ClusterIP 10.97.250.131 2181/TCP 2m NAME READY STATUS RESTARTS AGE empire-backup-6f4567d5fd-gcrvg 1/1 Running 0 2m empire-hq-59475b4b64-mrdww 1/1 Running 0 2m empire-outpost-8888-78dffd49fb-tnnhf 1/1 Running 0 2m empire-outpost-9999-7dd9fc5f5b-xp6jw 1/1 Running 0 2m kafka-broker-b874c78fd-jdwqf 1/1 Running 0 2m zookeeper-85f64b8cd4-nprck 1/1 Running 0 2m Setup Client Terminals ====================== First we will | https://github.com/cilium/cilium/blob/main//Documentation/security/kafka.rst | main | cilium | [
-0.04046643152832985,
0.02278849668800831,
-0.09388122707605362,
0.0038697971031069756,
0.09907074272632599,
-0.03811698779463768,
-0.05920884758234024,
0.012959090992808342,
0.008976109325885773,
0.003893120214343071,
0.03654778003692627,
-0.07750915735960007,
0.033789243549108505,
-0.052... | 0.138454 |
10.96.0.1 443/TCP 10m zook ClusterIP 10.97.250.131 2181/TCP 2m NAME READY STATUS RESTARTS AGE empire-backup-6f4567d5fd-gcrvg 1/1 Running 0 2m empire-hq-59475b4b64-mrdww 1/1 Running 0 2m empire-outpost-8888-78dffd49fb-tnnhf 1/1 Running 0 2m empire-outpost-9999-7dd9fc5f5b-xp6jw 1/1 Running 0 2m kafka-broker-b874c78fd-jdwqf 1/1 Running 0 2m zookeeper-85f64b8cd4-nprck 1/1 Running 0 2m Setup Client Terminals ====================== First we will open a set of windows to represent the different Kafka clients discussed above. For consistency, we recommend opening them in the pattern shown in the image below, but this is optional. .. image:: images/cilium\_kafka\_gsg\_terminal\_layout.png In each window, use copy-paste to have each terminal provide a shell inside each pod. empire-hq terminal: .. code-block:: shell-session $ HQ\_POD=$(kubectl get pods -l app=empire-hq -o jsonpath='{.items[0].metadata.name}') && kubectl exec -it $HQ\_POD -- sh -c "PS1=\"empire-hq $\" /bin/bash" empire-backup terminal: .. code-block:: shell-session $ BACKUP\_POD=$(kubectl get pods -l app=empire-backup -o jsonpath='{.items[0].metadata.name}') && kubectl exec -it $BACKUP\_POD -- sh -c "PS1=\"empire-backup $\" /bin/bash" outpost-8888 terminal: .. code-block:: shell-session $ OUTPOST\_8888\_POD=$(kubectl get pods -l outpostid=8888 -o jsonpath='{.items[0].metadata.name}') && kubectl exec -it $OUTPOST\_8888\_POD -- sh -c "PS1=\"outpost-8888 $\" /bin/bash" outpost-9999 terminal: .. code-block:: shell-session $ OUTPOST\_9999\_POD=$(kubectl get pods -l outpostid=9999 -o jsonpath='{.items[0].metadata.name}') && kubectl exec -it $OUTPOST\_9999\_POD -- sh -c "PS1=\"outpost-9999 $\" /bin/bash" Test Basic Kafka Produce & Consume ================================== First, let's start the consumer clients listening to their respective Kafka topics. All of the consumer commands below will hang intentionally, waiting to print data they consume from the Kafka topic: In the \*empire-backup\* window, start listening on the top-secret \*deathstar-plans\* topic: .. code-block:: shell-session $ ./kafka-consume.sh --topic deathstar-plans In the \*outpost-8888\* window, start listening to \*empire-announcement\*: .. code-block:: shell-session $ ./kafka-consume.sh --topic empire-announce Do the same in the \*outpost-9999\* window: .. code-block:: shell-session $ ./kafka-consume.sh --topic empire-announce Now from the \*empire-hq\*, first produce a message to the \*empire-announce\* topic: .. code-block:: shell-session $ echo "Happy 40th Birthday to General Tagge" | ./kafka-produce.sh --topic empire-announce This message will be posted to the \*empire-announce\* topic, and shows up in both the \*outpost-8888\* and \*outpost-9999\* windows who consume that topic. It will not show up in \*empire-backup\*. \*empire-hq\* can also post a version of the top-secret deathstar plans to the \*deathstar-plans\* topic: .. code-block:: shell-session $ echo "deathstar reactor design v3" | ./kafka-produce.sh --topic deathstar-plans This message shows up in the \*empire-backup\* window, but not for the outposts. Congratulations, Kafka is working as expected :) The Danger of a Compromised Kafka Client ======================================== But what if a rebel spy gains access to any of the remote outposts that act as Kafka clients? Since every client has access to the Kafka broker on port 9092, it can do some bad stuff. For starters, the outpost container can actually switch roles from a consumer to a producer, sending "malicious" data to all other consumers on the topic. To prove this, kill the existing ``kafka-consume.sh`` command in the outpost-9999 window by typing control-C and instead run: .. code-block:: shell-session $ echo "Vader Booed at Empire Karaoke Party" | ./kafka-produce.sh --topic empire-announce Uh oh! Outpost-8888 and all of the other outposts in the empire have now received this fake announcement. But even more nasty from a security perspective is that the outpost container can access any topic on the kafka-broker. In the outpost-9999 container, run: .. code-block:: shell-session $ ./kafka-consume.sh --topic deathstar-plans "deathstar reactor design v3" We see that any outpost can actually access the secret deathstar plans. Now we know how the rebels got access to them! Securing Access to Kafka with Cilium ==================================== Obviously, it would be much more secure to limit each pod's access to the Kafka broker to be least privilege (i.e., only what is needed for the app to | https://github.com/cilium/cilium/blob/main//Documentation/security/kafka.rst | main | cilium | [
-0.016441764310002327,
-0.06034506484866142,
-0.061502788215875626,
-0.040448132902383804,
-0.06488774716854095,
0.028413359075784683,
-0.03557963669300079,
-0.033561985939741135,
-0.012925105169415474,
0.0090999286621809,
-0.017262371256947517,
-0.0779116079211235,
-0.03582308068871498,
-... | 0.048802 |
the secret deathstar plans. Now we know how the rebels got access to them! Securing Access to Kafka with Cilium ==================================== Obviously, it would be much more secure to limit each pod's access to the Kafka broker to be least privilege (i.e., only what is needed for the app to operate correctly and nothing more). We can do that with the following Cilium security policy. As with Cilium HTTP policies, we can write policies that identify pods by labels, and then limit the traffic in/out of this pod. In this case, we'll create a policy that identifies the exact traffic that should be allowed to reach the Kafka broker, and deny the rest. As an example, a policy could limit containers with label \*app=empire-outpost\* to only be able to consume topic \*empire-announce\*, but would block any attempt by a compromised container (e.g., empire-outpost-9999) from producing to \*empire-announce\* or consuming from \*deathstar-plans\*. .. image:: images/cilium\_kafka\_gsg\_attack.png Here is the \*CiliumNetworkPolicy\* rule that limits access of pods with label \*app=empire-outpost\* to only consume on topic \*empire-announce\*: .. literalinclude:: ../../examples/policies/getting-started/kafka.yaml A \*CiliumNetworkPolicy\* contains a list of rules that define allowed requests, meaning that requests that do not match any rules are denied as invalid. The above rule applies to inbound (i.e., "ingress") connections to kafka-broker pods (as indicated by "app: kafka" in the "endpointSelector" section). The rule will apply to connections from pods with label "app: empire-outpost" as indicated by the "fromEndpoints" section. The rule explicitly matches Kafka connections destined to TCP 9092, and allows consume/produce actions on various topics of interest. For example we are allowing \*consume\* from topic \*empire-announce\* in this case. The full policy adds two additional rules that permit the legitimate "produce" (topic \*empire-announce\* and topic \*deathstar-plans\*) from \*empire-hq\* and the legitimate consume (topic = "deathstar-plans") from \*empire-backup\*. The full policy can be reviewed by opening the URL in the command below in a browser. Apply this Kafka-aware network security policy using ``kubectl`` in the main window: .. parsed-literal:: $ kubectl create -f \ |SCM\_WEB|\/examples/kubernetes-kafka/kafka-sw-security-policy.yaml If we then again try to produce a message from outpost-9999 to \*empire-annnounce\*, it is denied. Type control-c and then run: .. code-block:: shell-session $ echo "Vader Trips on His Own Cape" | ./kafka-produce.sh --topic empire-announce >>[2018-04-10 23:50:34,638] ERROR Error when sending message to topic empire-announce with key: null, value: 27 bytes with error: (org.apache.kafka.clients.producer.internals.ErrorLoggingCallback) org.apache.kafka.common.errors.TopicAuthorizationException: Not authorized to access topics: [empire-announce] This is because the policy does not allow messages with role = "produce" for topic "empire-announce" from containers with label app = empire-outpost. Its worth noting that we don't simply drop the message (which could easily be confused with a network error), but rather we respond with the Kafka access denied error (similar to how HTTP would return an error code of 403 unauthorized). Likewise, if the outpost container ever tries to consume from topic \*deathstar-plans\*, it is denied, as role = consume is only allowed for topic \*empire-announce\*. To test, from the outpost-9999 terminal, run: .. code-block:: shell-session $./kafka-consume.sh --topic deathstar-plans [2018-04-10 23:51:12,956] WARN Error while fetching metadata with correlation id 2 : {deathstar-plans=TOPIC\_AUTHORIZATION\_FAILED} (org.apache.kafka.clients.NetworkClient) This is blocked as well, thanks to the Cilium network policy. Imagine how different things would have been if the empire had been using Cilium from the beginning! Clean Up ======== You have now installed Cilium, deployed a demo app, and tested both L7 Kafka-aware network security policies. To clean up, run: .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-kafka/kafka-sw-app.yaml $ kubectl delete cnp secure-empire-kafka After this, you can re-run the tutorial from Step 1. | https://github.com/cilium/cilium/blob/main//Documentation/security/kafka.rst | main | cilium | [
0.017298273742198944,
0.053866226226091385,
-0.08517327904701233,
-0.07831893116235733,
0.003208475187420845,
-0.0028524533845484257,
0.04527871310710907,
0.0008349064155481756,
0.03613635152578354,
0.08067841827869415,
-0.005822095554322004,
-0.04196259006857872,
0.02781384065747261,
-0.0... | 0.09944 |
a demo app, and tested both L7 Kafka-aware network security policies. To clean up, run: .. parsed-literal:: $ kubectl delete -f \ |SCM\_WEB|\/examples/kubernetes-kafka/kafka-sw-app.yaml $ kubectl delete cnp secure-empire-kafka After this, you can re-run the tutorial from Step 1. | https://github.com/cilium/cilium/blob/main//Documentation/security/kafka.rst | main | cilium | [
-0.015585849992930889,
0.008511311374604702,
-0.04460284113883972,
-0.032458577305078506,
0.020232146605849266,
0.015899550169706345,
-0.04398728534579277,
-0.10112398862838745,
0.0039778659120202065,
0.08335084468126297,
-0.0050050909630954266,
-0.06629688292741776,
-0.011401998810470104,
... | 0.069807 |
.. only:: not (epub or latex or html) WARNING: You are looking at unreleased Cilium documentation. Please use the official rendered version released here: https://docs.cilium.io Threat Model ============ This section presents a threat model for Cilium. This threat model allows interested parties to understand: - security-specific implications of Cilium's architecture - controls that are in place to secure data flowing through Cilium's various components - recommended controls for running Cilium in a production environment Scope and Prerequisites ----------------------- This threat model considers the possible attacks that could affect an up-to-date version of Cilium running in a production environment; it will be refreshed when there are significant changes to Cilium's architecture or security posture. This model does not consider supply-chain attacks, such as attacks where a malicious contributor is able to intentionally inject vulnerable code into Cilium. For users who are concerned about supply-chain attacks, Cilium's `security audit`\_ assessed Cilium's supply chain controls against `the SLSA framework`\_. In order to understand the following threat model, readers will need familiarity with basic Kubernetes concepts, as well as a high-level understanding of Cilium's :ref:`architecture and components`. .. \_security audit: https://github.com/cilium/cilium.io/blob/main/Security-Reports/CiliumSecurityAudit2022.pdf .. \_the SLSA framework: https://slsa.dev/ Methodology ----------- This threat model considers eight different types of threat actors, placed at different parts of a typical deployment stack. We will primarily use Kubernetes as an example but the threat model remains accurate if deployed with other orchestration systems, or when running Cilium outside of Kubernetes. The attackers will have different levels of initial privileges, giving us a broad overview of the security guarantees that Cilium can provide depending on the nature of the threat and the extent of a previous compromise. For each threat actor, this guide uses the `the STRIDE methodology`\_ to assess likely attacks. Where one attack type in the STRIDE set can lead to others (for example, tampering leading to denial of service), we have described the attack path under the most impactful attack type. For the potential attacks that we identify, we recommend controls that can be used to reduce the risk of the identified attacks compromising a cluster. Applying the recommended controls is strongly advised in order to run Cilium securely in production. .. \_the STRIDE methodology: https://en.wikipedia.org/wiki/STRIDE\_(security) Reference Architecture ---------------------- For ease of understanding, consider a single Kubernetes cluster running Cilium, as illustrated below: .. image:: images/cilium\_threat\_model\_reference\_architecture.png The Threat Surface ~~~~~~~~~~~~~~~~~~ In the above scenario, the aim of Cilium's security controls is to ensure that all the components of the Cilium platform are operating correctly, to the extent possible given the abilities of the threat actor that Cilium is faced with. The key components that need to be protected are: - the Cilium agent running on a node, either as a Kubernetes pod, a host process, or as an entire virtual machine - Cilium state (either stored via CRDs or via an external key-value store like etcd) - eBPF programs loaded by Cilium into the kernel - network packets managed by Cilium - observability data collected by Cilium and stored by Hubble The Threat Model ---------------- For each type of attacker, we consider the plausible types of attacks available to them, how Cilium can be used to protect against these attacks, as well as the security controls that Cilium provides. For attacks which might arise as a consequence of the high level of privileges required by Cilium, we also suggest mitigations that users should apply to secure their environments. .. \_kubernetes-workload-attacker: Kubernetes Workload Attacker ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ For the first scenario, consider an attacker who has been able to gain access to a Kubernetes pod, and is now able to run | https://github.com/cilium/cilium/blob/main//Documentation/security/threat-model.rst | main | cilium | [
-0.013026203028857708,
0.01242824923247099,
-0.062331683933734894,
0.016178511083126068,
0.07413943856954575,
-0.04985278472304344,
-0.05069480091333389,
0.04269721731543541,
0.032241128385066986,
-0.026541847735643387,
0.07130222767591476,
-0.033799730241298676,
0.06389569491147995,
-0.02... | 0.194478 |
the high level of privileges required by Cilium, we also suggest mitigations that users should apply to secure their environments. .. \_kubernetes-workload-attacker: Kubernetes Workload Attacker ~~~~~~~~~~~~~~~~~~~~~~~~~~~~ For the first scenario, consider an attacker who has been able to gain access to a Kubernetes pod, and is now able to run arbitrary code inside a container. This could occur, for example, if a vulnerable service is exposed externally to a network. In this case, let us also assume that the compromised pod does not have any elevated privileges (in Kubernetes or on the host) or direct access to host files. .. image:: images/cilium\_threat\_model\_workload.png In this scenario, there is no potential for compromise of the Cilium stack; in fact, Cilium provides several features that would allow users to limit the scope of such an attack: .. rst-class:: wrapped-table +-----------------+---------------------+--------------------------------+ | Threat surface | Identified STRIDE | Cilium security benefits | | | threats | | +=================+=====================+================================+ | Cilium agent | Potential denial of | Cilium can enforce | | | service if the | `bandwidth limitations`\_ | | | compromised | on pods to limit the network | | | | resource utilization. | | | Kubernetes workload | | | | does not have | | | | defined resource | | | | limits. | | +-----------------+---------------------+--------------------------------+ | Cilium | None | | | configuration | | | +-----------------+---------------------+--------------------------------+ | Cilium eBPF | None | | | programs | | | +-----------------+---------------------+--------------------------------+ | Network data | None | - Cilium's network policy can | | | | be used to provide | | | | least-privilege isolation | | | | between Kubernetes | | | | workloads, and between | | | | Kubernetes workloads and | | | | "external" endpoints running | | | | outside the Kubernetes | | | | cluster, or running on the | | | | Kubernetes worker nodes. | | | | Users should ideally define | | | | specific allow rules that | | | | only permit expected | | | | communication between | | | | services. | | | | - Cilium's network | | | | connectivity will prevent an | | | | attacker from observing the | | | | traffic intended for other | | | | workloads, or sending | | | | traffic that "spoofs" the | | | | identity of another pod, | | | | even if transparent | | | | encryption is not in use. | | | | Pods cannot send traffic | | | | that "spoofs" other pods due | | | | to limits on the use of | | | | source IPs and limits on | | | | sending tunneled traffic. | +-----------------+---------------------+--------------------------------+ | Observability | None | Cilium's Hubble flow-event | | data | | observability can be used to | | | | provide reliable audit of | | | | the attacker's L3/L4 and L7 | | | | network connectivity. | +-----------------+---------------------+--------------------------------+ .. \_bandwidth limitations: https://docs.cilium.io/en/stable/network/kubernetes/bandwidth-manager/ Recommended Controls ^^^^^^^^^^^^^^^^^^^^ - Kubernetes workloads should have `defined resource limits`\_. This will help in ensuring that Cilium is not starved of resources due to a misbehaving deployment in a cluster. - Cilium can be given prioritized access to system resources either via Kubernetes, cgroups, or other controls. - Runtime security solutions such as `Tetragon`\_ should be deployed to ensure that container compromises can be detected as they occur. .. \_defined resource limits: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ .. \_Tetragon: https://github.com/cilium/tetragon .. \_limited-privilege-host-attacker: Limited-privilege Host Attacker ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In this scenario, the | https://github.com/cilium/cilium/blob/main//Documentation/security/threat-model.rst | main | cilium | [
-0.002524424809962511,
0.05918101221323013,
-0.03433127701282501,
-0.009037160314619541,
0.00760072423145175,
-0.06662143766880035,
0.03455900773406029,
0.07301095873117447,
0.047140613198280334,
0.023262856528162956,
0.007419360801577568,
-0.020483341068029404,
0.026471680030226707,
0.024... | 0.215374 |
access to system resources either via Kubernetes, cgroups, or other controls. - Runtime security solutions such as `Tetragon`\_ should be deployed to ensure that container compromises can be detected as they occur. .. \_defined resource limits: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ .. \_Tetragon: https://github.com/cilium/tetragon .. \_limited-privilege-host-attacker: Limited-privilege Host Attacker ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In this scenario, the attacker is someone with the ability to run arbitrary code with direct access to the host PID or network namespace (or both), but without "root"-equivalent privileges that would allow them to disable Cilium components or undermine the eBPF and other kernel state Cilium relies on. This level of access could exist for a variety of reasons, including: - Pods or other containers running in the host PID or network namespace, but without "root" privileges or capabilities. This includes ``hostNetwork: true`` and ``hostPID: true`` containers. - Non-"root" SSH or other console access to a node. - A containerized workload that has "escaped" the container namespace but as a non-privileged user. .. image:: images/cilium\_threat\_model\_non\_privileged.png In this case, an attacker would be able to bypass some of Cilium's network controls, as described below: .. rst-class:: wrapped-table +-----------------+-------------------------+----------------------------+ | \*\*Threat | \*\*Identified STRIDE | \*\*Cilium security | | surface\*\* | threats\*\* | benefits\*\* | +=================+=========================+============================+ | Cilium agent | - If the non-privileged | | | | attacker is able to | | | | access the container | | | | runtime and Cilium is | | | | running as a | | | | container, the | | | | attacker will be able | | | | to tamper with the | | | | Cilium agent running | | | | on the node. | | | | - Denial of service is | | | | also possible via | | | | spawning workloads | | | | directly on the host. | | +-----------------+-------------------------+----------------------------+ | Cilium | Same as for the Cilium | | | configuration | agent. | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | +-----------------+-------------------------+----------------------------+ | Cilium eBPF | Same as for the Cilium | | | programs | agent. | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | +-----------------+-------------------------+----------------------------+ | Network data | Elevation of | Cilium's network | | | privilege: traffic | connectivity will prevent | | | sent by the attacker | an attacker from observing | | | will no longer be | the traffic intended for | | | subject to Kubernetes | other workloads, or | | | or | sending traffic that | | | container-networked | spoofs the identity of | | | Cilium network | another pod, even if | | | policies. | transparent encryption is | | | :ref:`Host-networked | not in use. | | | Cilium | | | | policies | | | | ` | | | | will continue to | | | | apply. Other traffic | | | | within the cluster | | | | remains unaffected. | | +-----------------+-------------------------+----------------------------+ | Observability | None | Cilium's Hubble flow-event | | data | | observability can be used | | | | to provide reliable audit | | | | of the attacker's L3/L4 | | | | and L7 network | | | | connectivity. Traffic sent | | | | by the attacker will be | | | | attributed to | https://github.com/cilium/cilium/blob/main//Documentation/security/threat-model.rst | main | cilium | [
-0.034246597439050674,
0.08876721560955048,
-0.053758759051561356,
-0.02129276841878891,
0.015428640879690647,
-0.03359227254986763,
0.029253020882606506,
0.036212410777807236,
0.10514416545629501,
0.024395441636443138,
0.006930972915142775,
-0.07782067358493805,
0.02105793170630932,
-0.03... | 0.20423 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.