Radix Config
In order for Radix to configure your application it needs a configuration file. By default, it is expected to be located in the root of the application repository, has a name radixconfig.yaml
and be in YAML or JSON format - in either case, it must have the .yaml
or .yml
extension (the name and extension should be exactly same as for the file in the GitHub repository). The name of the file and its location in the repository can be different. It can also be changed later on the Radix web-console configuration page for the application. Read more in the monorepo guide.
Radix only reads radixconfig.yaml
from the branch we set as the Config Branch
in the application registration form. If the file is changed in other branches, those changes will be ignored.
The basic format of the file is this; the configuration keys are explained in the Reference section below:
apiVersion: radix.equinor.com/v1
kind: RadixApplication
metadata: ...
spec: ...
You can enable auto-completion and validation for radixconfig.yaml
in your code editor. Read more about it here.
Reference
name
apiVersion: radix.equinor.com/v1
kind: RadixApplication
metadata:
name: myapp
spec:
name
needs to match the name given in when registering an application. Only lowercase characters are allowed. If the name supplied in the configuration contains uppercase characters, a warning will be logged and the name will be automatically converted to lowercase.
build
spec:
build:
useBuildKit: true
useBuildCache: true
secrets:
- SECRET_1
- SECRET_2
variables:
VAR1: val1
CONNECTION_STRING: "Provider=MySQLProv;Data Source=mydb;"
The build
section of the spec contains configuration used during the build process of the components and jobs.
useBuildKit
useBuildKit
- (optional, default false
) builds components and jobs using Buildah. This option provides several benefits over the default Radix build engine:
- Secure handling of build secrets.
- Caching support that can reduce build time, see
useBuildCache
. - Use images from protected container registries defined in
privateImageHubs
, in the Dockerfile'sFROM
instructions.
Buildah requires the image name in a Dockerfile's FROM
instructions to be fully qualified, e.g. FROM docker.io/library/golang:1.22
instead of FROM golang:1.22
. If you are using the style bitnami/golang:1.22
you should use docker.io/bitnami/golang:1.22
The build step will fail with the following error an image name is not fully qualified:
Error: creating build container: short-name resolution enforced but cannot prompt without a TTY
useBuildCache
useBuildCache
- (optional, defaults to true
) pushes all layers to cache, and uses it in future builds when possible. Requires useBuildKit
to be enabled. Internally we set --cache-to
, --cache-from
and --layers
in Buildah. Read more at Buildahs Documentation
This option can be overridden in the Radix CLI command rx create pipeline-job build-deploy
with an argument --use-build-cache=true|false
Make sure you never store secrets or confidential information in any intermitent layer, multistage image, or in your final container image.
secrets
secrets
- (optional) Defines secrets to be used in Dockerfiles or sub-pipelines. Secrets values must be set in Radix Web Console. build-deploy
jobs will fail if not all secret values are set.
- When an option
useBuildKit: false
, to ensure that multiline build secrets are handled correct by the build, all build secrets are passed asARG
-s during container build, base-64 encoded (they need to be decoded before use). - When an option
useBuildKit: true
, build secrets are not available asARG
-s during container build, but they can be mounted as files. Secret values are not base-64 encoded in these files.
Read the build secrets guide to see how to use build secrets in a Dockerfile.
variables
spec:
build:
subPipeline:
variables:
VAR1: value1
VAR2: value2
variables
- (optional, available only in sub-pipelines, they are ignored for sub-pipeline if the subPipeline option exists) environment variables names and values, provided for all build Radix environments in sub-pipelines. These common environment variables are overridden by environment-specific environment variables with the same names.
subPipeline
subPipeline
- (optional, available only in sub-pipelines) configuration of sub-pipeline options. When it exists, variables are not used in sub-pipeline.
variables
spec:
build:
subPipeline:
variables:
VAR1: value1
VAR2: value2
Sub-pipeline environment variables names and values, provided for all build Radix environments in sub-pipelines. These common environment variables can be overridden by environment-specific environment variables with the same names.
identity
spec:
build:
subPipeline:
identity:
azure:
clientId: 12345678-a263-abcd-8993-683cc6123456
When identity.azure.clientId
option is set, the environment variable AZURE_CLIENT_ID
with its value is automatically added to the running pipeline, and it can be used in this pipeline tasks. Read more about the identity in the component identity option and about using it in the sub-pipeline in the Pipeline with Azure workload identity example.
environments
spec:
environments:
- name: dev
build:
from: master
- name: prod
build:
from: release
The environments
section of the spec lists the environments for the application.
name
The name of the environment. Can be dev
, qa
, production
etc.
build
from
specifies which branch each environment will build from. If from
is not specified for the environment, no automatic builds or deployments will be created. This configuration is useful for a promotion-based workflow.
Wildcard branch mapping is also support, using *
and ?
. Examples:
feature/*
feature-?
hotfix/**/*
variables
- environment variable names and values (currently available only in sub-pipelines), provided for specific environments. They override common environment variables with the same names, if specified in the spec.build.variables
.
A text input field, will be available to put a full branch name for a build environment.
Example:
spec:
build:
variables:
VAR1: val1
CONNECTION_STRING: "Provider=MySQLProv;Data Source=prodDb;"
environments:
- name: dev
build:
from: main
variables:
VAR1: "val1-for-dev" #overrides common env-var VAR1 in the "dev" external pipeline
CONNECTION_STRING: "Provider=MySQLProv;Data Source=devDb;" #overrides common env-var CONNECTION_STRING in the "dev" custom sub-pipeline
- name: prod
build:
from: release
variables:
VAR3: val3 #env-var VAR3 only exists in the "prod" custom sub-pipeline, in addition to common VAR1 and CONNECTION_STRING
egress
spec:
environments:
- name: dev
build:
from: master
egress:
allowRadix: true
rules:
- destinations:
- "143.97.5.5/32"
- "143.97.6.1/32"
ports:
- port: 443
protocol: TCP
- name: prod
build:
from: release
Specify egress
with settings for which egress traffic is allowed from all components and jobs in the environment.
allowRadix
can be set to true
or false
to allow or deny traffic to other applications in Radix. The default value is false
.
rules
can be defined with a list of legal destinations
and ports
for egress traffic. Each entry in destinations
must be a string representing a valid IPv4 mask. Each entry in ports
must be an object with a valid TCP/UDP port
number and protocol
equal to either "TCP" or "UDP". If one or more egress rules are defined, any traffic not allowed by the egress rules will be blocked. If no egress rules are defined, all traffic is allowed.
See the egress configuration guide for usage patterns and tips and tricks.
If an environment
has defined the egress
field, all traffic is blocked by default. If egress
is not defined, all traffic is allowed.
If your application uses a custom OAuth2 implementation, outbound access to Microsoft authentication endpoints must be allowed. See allow traffic for OAuth2.
subPipeline
subPipeline
- (optional, available only in sub-pipelines) configuration of sub-pipeline options for specific environment.
- It can override common subPipeline or combine with it (if present) for a specific environment.
- It can remove all common environment variables and the common identity (if present) with
{}
(empty object) for a specific environment
spec:
environments:
- name: dev
subPipeline: {}
variables
spec:
environments:
- name: dev
subPipeline:
variables:
VAR1: value1
VAR2: value2
Sub-pipeline environment variables names and values, provided for specific build Radix environment in sub-pipelines. These variables will be combined with common environment variables (if present).
identity
spec:
environments:
- name: dev
subPipeline:
identity:
azure:
clientId: 12345678-a263-abcd-8993-683cc6123456
The identity
section enables identity for a specific environment. Read mode about build identity.
- It can remove the common identity with
{}
empty object for a specific environment
spec:
environments:
- name: dev
subPipeline:
identity:
azure: {}
components
This is where you specify the various components for your application - it needs at least one. Each component needs a name
; this will be used for building the Docker images (appName-componentName). Source for the component can be; a folder in the repository, a dockerfile or an image.
src
spec:
components:
- name: frontend
src: frontend
ports:
- name: http
port: 8080
- name: backend
src: backend
ports:
- name: http
port: 5000
src
a folder, relative to the repository root, where the Dockerfile
for the component is located. The Dockerfile is used by the Build and deploy workflow of the Radix CI-CD pipeline to build a container image for the component. By default src
is .
- a root of the GitHub repository.
For Radix environment specific src
, refer to environmentConfig src.
When the image
option is set - src
option is ignored.
dockerfileName
spec:
components:
- name: frontend
dockerfileName: Dockerfile # Absolute path from repository root: /Dockerfile
ports:
- name: http
port: 8080
- name: backend
src: .
dockerfileName: backend/Dockerfile # Absolute path from repository root: /backend/Dockerfile
ports:
- name: http
port: 5000
- name: api
src: api
dockerfileName: "../Dockerfile" # Absolute path from repository root: /Dockerfile
ports:
- name: http
port: 5000
- name: web
src: web
dockerfileName: "subfolder/Dockerfile" # Absolute path from repository root: /web/subfolder/Dockerfile
ports:
- name: http
port: 5000
By default, Radix pipeline expects a docker file with a name Dockefile
in the component src
folder. If this file name needs to be different, it can be specified in the option dockerfileName
. The name can also contain a path relative to src
. See configuration examples above.
For Radix environment specific dockerfileName
, refer to environmentConfig image.
When the image
option is set - dockerfileName
option is ignored.
image
An alternative configuration of a component could be to use an existing container image, which can be specified in the option image
. When this option is set for a component - the component will be not build, but only deployed with this image. An example of such a configuration:
spec:
components:
- name: redis
image: redis:5.0-alpine
- name: swagger-ui
image: swaggerapi/swagger-ui
ports:
- name: http
port: 8080
publicPort: http
- When a container image is from the DockerHub repository, it is enough to specify only the image name. Examples:
image: redis:latest
image: redis:7.0.5
.- When an image is located in another container registry, the image name need to have the container registry URL. Example:
image: gcr.io/distroless/nodejs18-debian11
.image: gcr.io/distroless/nodejs18-debian11:latest
.
- When an image is not publicly available, it is required to provide an authentication information. Please read more about privateImageHubs option.
For Radix environment specific image
, refer to environmentConfig image.
ports
spec:
components:
- name: frontend
ports:
- name: http
port: 8000
- name: api
port: 8001
A component can optionally have one or several ports:
name
- internal name of a port, used as a reference within the radixconfig. It needs to be unique within the componentports
list.port
- numeric value of a port, in the range between 1024 and 65535. It needs to be unique within the componentports
list.
A component doesn't need to have ports. If it has at least one port, it has to respond to TCP or HTTP requests, sent to this port. Kubernetes readiness probe will regularly request the first port in the ports
list to ensure that the component can handle requests.
When a new component version is deployed, the probe waits until replicas of the new component version start responding to such requests, keeping them in the "Starting" state. When the new replicas respond to these requests, the rolling update will remove the replicas of the old component version
publicPort
spec:
components:
- name: frontend
publicPort: http
The publicPort
field of a component, if set to <PORT_NAME>
, is used to make the component accessible on the internet by generating a public endpoint. By default, the public endpoint can be accessed from all public IP addresses. You can restrict access to the public endpoints by configuring a list of IP address ranges in network.ingress.public.allow
, see network for more information.
A component without publicPort: <PORT_NAME>
can only be accessed from another component in the app. If specified, the <PORT_NAME>
should exist in the ports
field.
If no ports specified for a component, publicPort
should not be set.
monitoring
spec:
components:
- name: frontend
monitoring: true
When the monitoring
field is set to true
, is used to expose custom application metrics in the Radix monitoring dashboards. It is expected that the component provides a /metrics
endpoint: this will be queried periodically (every five seconds) by an instance of Prometheus running within Radix. General metrics, such as resource usage, will always be available in monitors, regardless of this being set.
monitoringConfig
spec:
components:
- name: frontend
ports:
- name: http
port: 8000
- name: metrics
port: 1234
monitoringConfig:
portName: metrics
path: /my-metrics
The monitoringConfig
field of a component can be used to override the default port and/or path used for monitoring
. Both fields are optional and are by default set to use the first available port and the path as /metrics
.
If overriding portName
it will have to match one of the defined ports in the component.
horizontalScaling
Scale your components replicas up and down based on resources or external metrics like CRON or Azure Service Bus.
If you need any other trigger types, find the list of available triggers at https://keda.sh/docs/latest/scalers/ and open a feature request on https://github.com/equinor/radix
You can override horizontalScaling in your environments and we will merge minReplicas
, maxReplicas
, pollingInterval
and cooldownPeriod
. If any triggers are defined in the environment, they will replace all triggers on the component level.
Read more about polling intervall and cooldown period in KEDAs documentation here
spec:
components:
- name: backend
horizontalScaling:
minReplicas: 0 # defaults to 1 if not set
maxReplicas: 6
# pollingInterval: 30 # Default
# cooldownPeriod: 300 # Default
triggers:
# Cpu/Memory triggers will scale up/down so the average usage
# is 85% of requested Cpu across all pods,
# or 75% of requested memory across all pods, whatever is highest.
- name: cpu
cpu:
value: 85
- name: memory
cpu:
value: 75
- name: cron
cron:
timezone: Europe/Oslo
start: 0 7 * * 1-5 # 07:00 Monday - Friday
end: 0 17 * * 1-5 # 17:00 Monday - Friday
desiredReplicas: 1
- name: azuresb
azureServiceBus:
namespace: <your servicebus namespace> #.servicebus.windows.net
# messageCount: 5
# activationMessageCount: 0
queueName: main
# or use TopicName/subscriptionName:
# topicName: my-topic
# subscriptionName: my-subscription
authentication: # Currently we only support Workload Identity
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000000
# Deperecated: legacy resources will be rewritten as triggers by Radix.
# - It is not allowed to mix resources with triggers.
# resources:
# memory:
# averageUtilization: 75
# cpu:
# averageUtilization: 85
The horizontalScaling
field is used for enabling automatic scaling of the component. This field is optional, and if set, it will override the replicas
value of the component. If no triggers are defined, we will configure a default CPU trigger with a target of 80% average usage.
One exception is when the replicas
value is set to 0
(i.e. the component is stopped), the horizontalScaling
config will not be used.
The previous resources
block have been replaced by triggers
.
azureServiceBus
trigger
Take a look here github.com/equinor/radix-sample-keda for a sample implementation that runs on Radix.
Azure Service Bus supports either a queueName
, or a topicName
and subscriptionName
. You can also select the target average messageCount
(defaults to 5), and activationMessageCount
(defaults to 0).
To authenticate Keda for scaling, you must provide a clientId to a managed identity, that contains a federeated credential with these properties:
Federated credential scenario: Kubernetes Service Account
# Find real and current value here: https://console.radix.equinor.com/about (CLUSTER_OIDC_ISSUER_URL),
# it will change in the future, we will post details in the slack channel #omnia-radix when it must be changed.
Cluster Issuer URL: https://northeurope.oic.prod-aks.azure.com/00000000-0000-0000-0000-000000000000/00000000-0000-0000-0000-000000000000/
Namespace: keda
Service Account: keda-operator
# ⚠️ When you give Keda access to your Service Bus, any other Radix app can scale their app based on your queue.
# We are hoping on improving this - https://github.com/kedacore/keda/issues/5630
healthChecks
By default Radix configures a TCP Readiness probe that tells the platform your component is ready to accept traffic as soon as the port is opened. Radix support Readiness, Liveness and Startup Probes.
If any custom probes are configured, Radix will not include the default readiness probe and you should configure it yourself.
A HTTP Get probe for readiness probe is usually easer to manage than a TCP probe, and the probe should let Radix now when your component and your dependencies are ready for traffic.
TCP Probes only checks if the port is open or closed. This way its hard to explicitly open or close the port when your component or dependencies are unavailable after startup. This could also disrupt regular requests that might not be affected.
spec:
components:
- name: backend
healthChecks:
startupProbe:
tcpSocket:
port: 8000
livenessProbe:
httpGet:
port: 8000
path: /healthz/live
readinessProbe:
successThreshold: 1
periodSeconds: 30
failureThreshold: 3
initialDelaySeconds: 10
timeoutSeconds: 5
httpGet:
port: 8000
path: /healthz/ready # Component is Ready when this endpoint returns a successful status code (2xx or 3xx)
All probes have all settings (except for successThreshold that is not available for liveness and startup probes). We support HTTP, TCP, Exec and GRPC probes.
environmentConfig
can also override individual startup, liveness or readiness-probes (but will not merge probe specific config).
Incorrectly configured probes can lead to premature restarts and will affect your uptime.
Read more about probes here: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-startup-probes/
imageTagName
components:
- name: backend
image: docker.pkg.github.com/equinor/myapp/backend:{imageTagName}
imageTagName: master-latest
The imageTagName
allows for flexible configuration of fixed images, built outside of Radix. It can be configured with separate tag for each environment.
See this guide on how make use of imageTagName
in a deploy-only scenario.
volumeMounts
spec:
components:
- name: backend
volumeMounts:
- name: volume-name
path: /path/in/container/to/mount/to
blobfuse2:
container: container-name
uid: 1000
- name: temp-volume-name
path: /another/path/in/container/to/mount/to
emptyDir:
sizeLimit: 10M
The volumeMounts
field configures volume mounts within the running component.
main
settings
name
- the name of the volume. Unique withinvolumeMounts
list of a componentpath
- the folder inside the running container, where the external storage is mounted.emptyDir
- mounts a read-write empty volume in the container.blobfuse2
- mount a container from blob in Azure storage account. Uses CSI Azure blob storage driver. Replaces typesblob
andazure-blob
for obsolete drivers.
emptyDir
and blobfuse2
are mutually exclusive.
emptyDir
settings
sizeLimit
- The maxiumum capacity for the volume.
An emptyDir
volume mounts a temporary writable volume in the container. Data in an emptyDir
volume is safe across container crashes for component replicas, but is lost if a job container crashes and restarts. When a component replica is deleted for any reason, the data in emptyDir
is removed permanently.
emptyDir
volumes are useful when readOnlyFileSystem
is set to true
.
- scratch space, such as for a disk-based merge sort
- checkpointing a long computation for recovery from crashes
- holding files that a content-manager container fetches while a webserver container serves the data
blobfuse2
settings
-
protocol
- (optional) a protocol, supported by the BlobFuse2. Currently, supportsfuse2
(default) andnfs
. -
container
- name of the blob container. -
uid
and/orgid
- User ID and/or group ID (numbers) of a mounted volume owner. It is a User ID and Group ID of a user in the running container within component replicas. Usually a user, which is a member of one or multiple groups, is specified in theDockerfile
for the component with commandUSER
. Read more details about specifying user withinDockerfile
. It is recommended to use because Blobfuse driver do not honor fsGroup securityContext settings. -
useAdls
- (optional) enables blobfuse to access Azure DataLake storage account. When set to false, blobfuse will access Azure Block Blob storage account, hierarchical file system is not supported. Defaultfalse
. This must be settrue
when HNS enabled account is mounted. -
streaming
- (optional) defines a file streaming. When it is turned on (it is by default), files, opened by a container in its volume mount are not cached on a node, but read directly from a blob storage. It is recommended to use. When it is turned off, files are cached on a node, but it may cause a problem with a limited node disk available space.Options for
streaming
enabled
- (optional) turn on/off a file streaming. Default istrue
There are optional settings to fine tune volumes.
Access to the Azure storage need to be set in secrets
for the component.
See this guide on how make use of volumeMounts
.
ingressConfiguration
spec:
components:
- name: frontend
ingressConfiguration:
- websocketfriendly
The ingressConfiguration
field of a component will add extra configuration by annotations to the Nginx ingress, useful for a particular scenario.
Note that the settings affect the connections with the public component, not between a public and a private component.
websocketfriendly
will change connection timeout to 1 hour for the component.stickysessions
will change load balancing of the ingress to route to a single replica.leastconnectedlb
will try to route connection to the replica with least amount of load
See this for more information on what annotations will be put on the ingress, given the configuration. See nginx documentation for details about how the native nginx load balancing methods work.
alwaysPullImageOnDeploy
spec:
components:
- name: api
image: docker.pkg.github.com/equinor/my-app/api:latest
alwaysPullImageOnDeploy: false
Only relevant for teams that uses another CI tool than Radix and static tags. See deploy-only for more information.
secrets
spec:
components:
- name: backend
secrets:
- DB_PASS
The secrets
key contains a list of names. Values for these can be set via the Radix Web Console (under each active component within an environment). Each secret must be set on all environments. Secrets are available in the component as environment variables; a component will not be able to start without the secret being set.
resources
(common)
spec:
components:
- name: backend
resources:
requests:
memory: "64Mi"
cpu: "50m"
limits:
cpu: "1000m"
The resources
section specifies how much CPU and memory each component needs, that are shared among all Radix environments in a component.
These common resources are overridden by environment-specific resources. The requested quota of memory and cpu must be below the limit.
If no memory limit is set, but a memory request is set, we will set the limit equal to the requested value. The opposite is also true, if a memory limit is set, but no requests, we will sett the requested memory equal to the memory limit.
Read more about memory resources and QoS.
More details about resources
and about default resources.
variables
(common)
spec:
components:
- name: backend
variables:
DB_NAME: my-db
The variables
key contains environment variable names and their values, that are shared among all Radix environments in a component. These common environment variables are overridden by environment-specific environment variables that have exactly same names.
Environment variables can be changed in Radix Console.
environmentConfig
The environmentConfig
section is to set environment-specific settings for each component.
src
When a component needs to be built with a different folder in a particular application environment, this folder can be specified in the src
option of the environmentConfig
section for this environment. If the component has an option image
specified on the component level, this image
option will be ignored and this component will be built for this environment with that specified src
. An example of such configuration:
spec:
components:
- name: frontend
src: "."
environmentConfig:
- environment: dev
src: "./frontend"
- name: backend
environmentConfig:
- environment: dev
src: "./backend"
- name: api
image: ghcr.io/my-repo/my-app/api:v1.10
environmentConfig:
- environment: dev
src: "./api"
In this example:
- The
frontend
component in thedev
environment will be built from thefrontend
folder of the GitHub repo, and from a root folder for other environments. Both docker-files expected to beDockerfile
asdockerfileName
option is not specified. - The
backend
component in thedev
environment will be built from thebackend
folder of the GitHub repo, and from a root folder (used by default) for other environments. Both docker-files expected to beDockerfile
asdockerfileName
option is not specified. - The
api
component in thedev
environment will be built from theapi
folder. This option in the environment overrides animage
option on the component level, which valueghcr.io/my-repo/my-app/api:v1.10
will be used to deploy this component in other environments. The docker-file expected to beDockerfile
asdockerfileName
option is not specified.
For shared src
across Radix environments, refer to common src.
When both image
and src
options are specified - the image
has higher priority, the src
option is ignored.
The src
option can be combined with dockerfileName
option.
dockerfileName
When a component needs to be built with a different docker-file in a particular application environment, this docker-file can be specified in the dockerfileName
option of the environmentConfig
section for this environment. If the component has an option image
specified on the component level, this image
option will be ignored and this component will be built for this environment with that specified dockerfileName
. An example of such configuration:
spec:
components:
- name: frontend
dockerfileName: Dockerfile
environmentConfig:
- environment: dev
dockerfileName: Dockerfile.dev
- name: backend
src: "./backend"
environmentConfig:
- environment: dev
dockerfileName: Dockerfile.dev
- name: api
image: ghcr.io/my-repo/my-app/api:v1.10
environmentConfig:
- environment: dev
dockerfileName: Dockerfile
In this example:
- The
frontend
component in thedev
environment will be built with theDockerfile.dev
docker-file, and with aDockerfile
docker-file for other environments. Both docker-files expected to be in the root of the GitHub repo assrc
option is not specified. - The
backend
component in thedev
environment will be built with theDockerfile.dev
docker-file, and with aDockerfile
docker-file (a name, used by default) for other environments. Both docker-files expected to be in thebackend
folder in the root of the GitHub repo. - The
api
component in thedev
environment will be built with theDockerfile
docker-file. This option in the environment overrides animage
option on the component level, which valueghcr.io/my-repo/my-app/api:v1.10
will be used to deploy this component in other environments. The docker-file expected to be in the root of the GitHub repo assrc
option is not specified.
For shared dockerfileName
across Radix environments, refer to common dockerfileName.
When both image
and dockerfileName
options are specified - the image
has higher priority, the dockerfileName
option is ignored.
The dockerfileName
option can be combined with src
option.
image
When a component needs a different docker image in a particular application environment, this image can be specified in the image
option of the environmentConfig
section for this environment. An example of such configuration:
spec:
components:
- name: redis
image: redis:5.0-alpine
environmentConfig:
- environment: qa
image: redis:7.2.4
- name: web-app
src: ./app
dockerfileName: Dockerfile.app
environmentConfig:
- environment: prod
image: ghcr.io/my-repo/my-app/web-app:v1.10
In this example:
- The
redis
component in theqa
environment will be run on the imageredis:7.2.4
, in other environments it will be run on the default imageredis:5.0-alpine
- The
web-app
component in theprod
environment will be run on the pre-build imageghcr.io/my-repo/my-app/web-app:v1.10
, in other environments it will be built from the source folder./app
and the docker-fileDockerfile.app
For shared image across Radix environments, refer to common image.
replicas
spec:
components:
- name: backend
environmentConfig:
- environment: prod
replicas: 2
replicas
can be used to horizontally scale the component. If replicas
is not set, it defaults to 1
. If replicas
is set to 0
, the component will not be deployed (i.e. stopped).
monitoring
spec:
components:
- name: frontend
environmentConfig:
- environment: prod
monitoring: true
When the monitoring
field of a component environment config is set to true
, is used to expose custom application metrics for the specific environment.
See monitoring for more information.
resources
spec:
components:
- name: frontend
environmentConfig:
- environment: prod
resources:
requests:
memory: "128Mi"
cpu: "100m"
limits:
cpu: "2000m"
The resources
section specifies how much CPU and memory each component needs, that are defined per Radix environment in a component. resources
is used to ensure that each component is allocated enough resources to run as it should. limits
describes the maximum amount of compute resources allowed. requests
describes the minimum amount of compute resources required. If requests
is omitted for a component it defaults to the settings in limits
. If limits
is omitted, its value defaults to an implementation-defined value. More info
The property limits.memory
cannot be explicitly set, it is set automatically with the same value as requests.memory
For shared resources across Radix environments, refer to common resources.
variables
spec:
components:
- name: backend
environmentConfig:
- environment: dev
variables:
DB_HOST: "db-dev"
DB_PORT: "1234"
- environment: prod
variables:
DB_HOST: "db-prod"
DB_PORT: "9876"
The variables
key contains environment variable names and their values, that are defined per Radix environment in a component. In addition to what is defined here, running containers will also have some environment variables automatically set by Radix.
For shared environment variables across Radix environments, refer to common environment variables.
horizontalScaling
spec:
components:
- name: backend
environmentConfig:
- environment: prod
horizontalScaling:
resources:
memory:
averageUtilization: 75
cpu:
averageUtilization: 85
minReplicas: 2
maxReplicas: 6
The horizontalScaling
field of a component environment config adds automatic scaling of the component in the environment, or it combines or overrides a component imageTagName
if it is defined.
healthChecks
Health check probes defined here will override probes defined on the component level.
spec:
components:
- name: backend
healthChecks:
startupProbe:
tcpSocket:
port: 8000
livenessProbe:
tcpSocket:
port: 8000
readinessProbe:
httpGet:
port: 8000
path: /health/ready
environmentConfig:
- environment: prod
healthChecks:
livenessProbe:
httpGet:
port: 8000
path: /health/alive
In this example, the prod environment will have 3 probes, a startup probe (TCP port 8000), a liveness check (HTTP /health/ready) and a readiness probe (HTTP /health/ready)
Please see more information at healthChecks
for detailed information and warnings.
imageTagName
components:
- name: backend
image: docker.pkg.github.com/equinor/myapp/backend:{imageTagName}
environmentConfig:
- environment: qa
imageTagName: master-latest
- environment: prod
imageTagName: release-39f1a082
The imageTagName
can be configured with separate tag for each environment. Environment imageTagName
overrides a component imageTagName
if it is also defined.
See imageTagName for a component for more information.
volumeMounts
spec:
components:
- name: backend
environmentConfig:
- environment: prod
volumeMounts:
- name: volume-name
path: /path/in/container/to/mount/to
blobfuse2:
container: container-name
uid: 1000
- name: temp-volume-name
path: /another/path/in/container/to/mount/to
emptyDir:
sizeLimit: 10M
The volumeMounts
field configures volume mounts within the component running in the specific environment. EnvironmentConfig volumeMounts
combine or override a component volumeMounts
if they are defined.
See volumeMounts for more information.
readOnlyFileSystem
spec:
components:
- name: backend
environmentConfig:
- environment: prod
readOnlyFileSystem: true|false
See readOnlyFileSystem for more information.
runtime
spec:
components:
- name: backend
environmentConfig:
- environment: prod
runtime:
architecture: amd64|arm64
See runtime for more information.
network
spec:
components:
- name: backend
environmentConfig:
- environment: prod
network:
ingress:
public:
allow:
- 100.1.1.1
- 100.2.2.2/30
See network for more information.
authentication
spec:
components:
- name: frontend
authentication:
clientCertificate: ...
oauth2: ...
environmentConfig:
- environment: dev
authentication:
clientCertificate: ...
oauth2: ...
The authentication
section can be used to configure an authentication option for either an entire component or a specific environment.
Note that the environment config will override the component config for that specific environment.
clientCertificate
clientCertificate:
verification: "optional_no_ca"
passCertificateToUpstream: true
clientCertificate
is a subsection of authentication and may be used to configure the Nginx Client Certificate Authentication.
Note that the Client Certificate configuration will be omitted if the component does not have a public port.
-
verification
Specifies type of verification of client certificates. Possible values are:off
: Don't request client certificates and don't do client certificate verification. (default)on
: Request a client certificate that must be signed by a certificate that is included in the secret key ca.crt of the secret specified bynginx.ingress.kubernetes.io/auth-tls-secret: secretName
. Failed certificate verification will result in a status code 400 (Bad Request).optional
: Do optional client certificate validation against the CAs from auth-tls-secret. The request fails with status code 400 (Bad Request) when a certificate is provided that is not signed by the CA. When no or an otherwise invalid certificate is provided, the request does not fail, but instead the verification result is sent to the upstream service.optional_no_ca
: Do optional client certificate validation, but do not fail the request when the client certificate is not signed by the CAs fromauth-tls-secret
. Certificate verification result is sent to the upstream service.
-
passCertificateToUpstream
Indicates if the received certificates should be passed or not to the upstream server in the header ssl-client-cert.verification
will have to be set to something other thanoff
for the certificate to be passed upstream. Possible values aretrue
orfalse
(default).
If verification
has been set to something other than off
or passCertificateToUpstream
is set to true
, a valid certificate will need to be applied in the Radix Console
for the affected environment(s). This can be found under Environments\[environmentName]\[componentName]\[componentName]-clientcertca
in the Radix Console
for your application.
oauth2
Configuration for adding OAuth2 authorization with OIDC to the component.
Common oauth2
settings can be configured at component level and/or in the component's environmentConfig
section. Properties configured in the environmentConfig
section override properties at component level. The component must also be configured with a publicPort.
When OAuth2 is configured for a component, Radix creates an OAuth2 service (using OAuth2 Proxy) to handle the OAuth2 authorization code flow, and to verify the authorization state of incoming requests to the component.
If no ports specified for a component, authentication.oauth2
should not be set.
The OAuth2 service handles incoming requests to the path /oauth2 (or the path defined in proxyPrefix) for all public DNS names configured for a component. Valid redirect URIs must be registered for the application registration in Azure AD, e.g. https://myapp.app.radix.equinor.com/oauth2/callback
.
See guide for more information.
oauth2:
clientId: ceeabc6f-93f0-41d2-9fe1-ca671b365f6f
scope: openid profile email offline_access 2b06f36c-5a78-45a6-9809-44816f59c1ff/user.read
setXAuthRequestHeaders: true
setAuthorizationHeader: true
proxyPrefix: /oauth2
oidc:
issuerUrl: https://login.microsoftonline.com/3aa4a235-b6e2-48d5-9195-7fcf05b459b0/v2.0
jwksUrl: https://login.microsoftonline.com/3aa4a235-b6e2-48d5-9195-7fcf05b459b0/discovery/v2.0/keys
skipDiscovery: false
insecureSkipVerifyNonce: false
loginUrl: https://login.microsoftonline.com/3aa4a235-b6e2-48d5-9195-7fcf05b459b0/oauth2/v2.0/authorize
redeemUrl: https://login.microsoftonline.com/3aa4a235-b6e2-48d5-9195-7fcf05b459b0/oauth2/v2.0/token
sessionStoreType: redis # redis or cookie
redisStore:
connectionUrl: rediss://app-session-store.redis.cache.windows.net:6380
cookieStore:
minimal: false
cookie:
name: _oauth2_proxy
expire: 168h0m0s
refresh: 60m0s
sameSite: lax
-
clientId
Required - The client ID of the application, e.g. the application ID of an application registration in Azure AD. -
scope
Optional. Default openid profile email - List of OIDC scopes and identity platform specific scopes. More information about scopes when using the Microsoft Identity Platform can be found here. -
setXAuthRequestHeaders
Optional. Default false - Adds claims from the access token to the X-Auth-Request-User, X-Auth-Request-Groups, X-Auth-Request-Email and X-Auth-Request-Preferred-Username request headers. The Access Token is added to the X-Auth-Request-Access-Token header. -
setAuthorizationHeader
Optional. Default false - Adds the OIDC ID Token in the Authorization: Bearer request header. -
proxyPrefix
Optional. Default /oauth2 - The root path that the OAuth2 proxy should be nested under. The OAuth2 proxy exposes various endpoints from this root path. -
oidc
OIDC configurationissuerUrl
Optional. Default <https://login.microsoftonline.com/3aa4a235-b6e2-48d5-9195-7fcf05b459b0/v2.0> - OIDC issuer URL. The default value is set to the Equinor Azure tenant ID. Read more about Microsoft OIDC issuer URLs here.jwksUrl
Optional - OIDC JWKS URL for token verification. Required ifskipDiscovery
is set to true.skipDiscovery
Optional. Default false - Skip automatic discovery OIDC endpoints.jwksURl
,loginUrl
andredeemUrl
must be specified ifskipDiscovery
is true.insecureSkipVerifyNonce
Optional. Default false. Skip verifying the OIDC ID Token's nonce claim. Should only be enabled with OIDC providers that does not support the nonce claim.
-
loginUrl
Optional - The authorization URL. Required ifskipDiscovery
is set to true. -
redeemUrl
Optional - The URL used to redeem authorization code and refresh token. Required ifskipDiscovery
is set to true. -
sessionStoreType
Optional. Default cookie. Allowed values: cookie, redis - Defines where session data shall be stored. -
redisStore
Redis session store configuration ifsessionStoreType
is redis.connectionUrl
Connection URL of redis server.
-
cookieStore
Cookie session store configuration ifsessionStoreType
is cookie.minimal
Optional. Default false. Strips ID token, access token and refresh token from session store cookies.setXAuthRequestHeaders
andsetAuthorizationHeader
must be set to false, andcookie.refresh
must be set to 0.
-
cookie
Session cookie configurationname
Optional. Default _oauth2_proxy. Name of the session cookie. IfsessionStoreType
is cookie, the ID token and access token is stored in cookies prefixed with this name.expire
Optional. Default 168h0m0s. Expire timeframe for session cookies. Controls the Expires cookie attribute.refresh
Optional. Default 60m0s. Refresh interval defines how often the OAuth2 service should redeem the refresh token to get a new access token. The session cookie's Expires is updated after refresh.sameSite
Optional. Default lax. The SameSite attribute for the session cookie.
See guide on how to configure OAuth2 authentication for a component.
enabled
Component can be disabled or enabled for all specific environment configurations.
spec:
components:
- name: backend
enabled: false
Disabled component is not deployed. When a disabling component is already deployed, this component is removed from the application environment on next deployment - this is an equivalent of removing of the component configuration from the radixconfig.yaml
.
When the option enabled
is not set, the component is enabled.
The component can be disabled in specific environment configurations.
spec:
components:
- name: backend
environmentConfig:
- environment: prod
enabled: false
Read more details in the guide.
Job components can be disabled similar way.
identity
spec:
components:
- name: backend
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000000
environmentConfig:
- environment: prod
identity: ...
The identity
section enables mounting of a JWT (JSON web token) that can be used as a federated credential with the OAuth2 client credentials flow to request an access token for an Azure AD application registration or managed identity.
The following environment variables are added to the replicas automatically when identity
is enabled:
- AZURE_AUTHORITY_HOST (
https://login.microsoftonline.com/
) - AZURE_CLIENT_ID (value from
clientId
in configuration, e.g.b96d264b-7053-4465-a4a7-32be5b0fec49
) - AZURE_FEDERATED_TOKEN_FILE (path to the file containg the JWT, e.g.
/var/run/secrets/azure/tokens/azure-identity-token
) - AZURE_TENANT_ID (e.g.
3aa4a235-b6e2-48d5-9195-7fcf05b459b0
)
identity
can be configured on the component/job level and/or per environment in the environmentConfig
section. Configuration in environmentConfig
overrides configuration on the component/job level.
See guide for more information.
readOnlyFileSystem
spec:
components:
- name: backend
readOnlyFileSystem: true|false
environmentConfig:
- environment: prod
readOnlyFileSystem: true|false
Mounts the container's root filesystem as read-only. Setting readOnlyFileSystem
in environmentConfig
overrides the value configured on component level. Defaults to false
if not specified.
Read-only filesystems will prevent the application from writing to disk. This is desirable in the event of an intrusion as the attacker will not be able to tamper with the filesystem or write foreign executables to disk. Without a writable filesystem the attack surface is dramatically reduced.
There may be a requirement for temporary files or local caching, in which case one or more writable emptyDir
volumes can be mounted.
runtime
spec:
components:
- name: backend
runtime:
architecture: amd64|arm64
environmentConfig:
- environment: prod
runtime:
architecture: amd64|arm64
The architecture
property in runtime
defines the CPU architecture a component or job should be built and deployed to. Valid values are amd64
and arm64
. The runtime
section can be configured on the component/job level and in environmentConfig
for a specific environment. environmentConfig
takes precedence over component/job level configuration. amd64
is used if neither is configured.
If you use the build and deploy
pipeline to build components or jobs, useBuildKit
must be enabled if at least one component or job is configured for arm64
. Radix will run the build steps on nodes with matching architecture, which in most cases mean that you do not have to change anything in the Dockerfile to support the configured architecture. This applies as long as the images defined in the Dockerfile's FROM <image>
supports this architecture.
For deploy-only components and jobs (with image
property set), make sure that the selected image supports the configured architecture. Many frequently used public images, like nginx-unprivileged and bitnami/redis, includes variants for both amd64
and arm64
in the same image. Radix (Kubernetes) will pull the appropriate variant based on the configured architecture.
network
spec:
environment:
- name: dev
- name: qa
- name: prod
components:
- name: backend
network:
ingress:
public:
proxyBodySize: 500m
proxyReadTimeout: 5
proxySendTimeout: 10
allow:
- 100.1.1.1
- 110.1.1.1/30
environmentConfig:
- environment: dev
network:
ingress:
proxyBodySize: 20m
proxyReadTimeout: 30
proxySendTimeout: 30
public:
allow: []
- environment: qa
network:
ingress:
public:
proxyBodySize: 100m
allow:
- 200.1.1.1
- 200.10.1.1
- environment: prod
network.ingress.public
contains settings used to control the behavior of public endpoints. These settings can be configured on the component level and/or in environmentConfig
for a specific environment. environmentConfig
takes precedence over component level configuration.
allow
: Defines a list of public IP addresses or CIDRs allowed to access the component's public endpoints. Settingallow
to an empty list allows access from all public IP addresses.
Note: Whenallow
is configured inenvironmentConfig
, it will overwrite any values defined on component level.proxyBodySize
: Sets the maximum allowed size of the client request body. Sizes can be specified in bytes, kilobytes (suffixes k and K), megabytes (suffixes m and M), or gigabytes (suffixes g and G), for example "1024", "64k", "32m" or "2g". If the size in a request exceeds the configured value, the 413 (Request Entity Too Large) error is returned to the client. Setting this value to "0" disables checking of client request body size. The default is 100m.proxyReadTimeout
: Defines a timeout, in seconds, for reading a response from the proxied server. The timeout is set only between two successive read operations, not for the transmission of the whole response. If the proxied server does not transmit anything within this time, the connection is closed. The default is 60 seconds.proxySendTimeout
: Defines a timeout, in seconds, for transmitting a request to the proxied server. The timeout is set only between two successive write operations, not for the transmission of the whole request. If the proxied server does not receive anything within this time, the connection is closed. The default is 60 seconds.
Setting proxyBodySize
to "0", or an unnecessary high value, can lead to instability/denial of service or increased cost, depending on how the request body is processed by the backend, e.g. when buffering to memory or storing the content to disk, either locally or remotely. Never set the value to "0" unless the backend component is configured to enforce a limit.
jobs
This is where you specify the various jobs for your application.
src
spec:
jobs:
- name: compute
src: compute
schedulerPort: 8000
- name: etl
src: etl
schedulerPort: 8000
See src for a component for more information.
dockerfileName
spec:
jobs:
- name: compute
dockerfileName: compute.Dockerfile
schedulerPort: 8000
See dockerfileName for a component for more information.
image
spec:
jobs:
- name: compute
image: privaterepodeleteme.azurecr.io/compute:latest
schedulerPort: 8000
See image for a component for more information.
schedulerPort
spec:
jobs:
- name: compute
schedulerPort: 8000
The port number that the job-scheduler will listen to for HTTP requests to manage jobs. schedulerPort is a required field.
In the example above, the URL for the compute job-scheduler is http://compute:8000
notifications
spec:
jobs:
- name: compute
notifications:
webhook: http://api:8080/monitor-batch-status
webhook
is an optional URL to the Radix application component or job component which will be called when any of the job-component's running jobs or batches changes states. Only changes are sent by POST method with a application/json
ContentType
in a batch event format. Read more
batchStatusRules
spec:
jobs:
- name: compute
batchStatusRules:
- condition: Any
operator: In
jobStatuses:
- Failed
batchStatus: Failed
- condition: All
operator: NotIn
jobStatuses:
- Waiting
- Active
- Running
batchStatus: Completed
batchStatusRules
- Optional rules to define batch statuses by their jobs statuses.
condition
-Any
,All
operator
-In
,NotIn
jobStatuses
-Waiting
,Active
,Running
,Succeeded
,Failed
,Stopped
batchStatus
-Waiting
,Active
,Running
,Succeeded
,Failed
,Stopping
,Stopped
,Completed
Rules are applied in the order from top to bottom in the rules list. When any rule matches, rules following it are ignored.
If batchStatusRules
are not defined or no rules match a batch status is set by following rules:
Waiting
- no jobs are startedActive
- any jobs are inActive
orRunning
stateCompleted
- no jobs are inWaiting
,Active
orRunning
states
Batch statuses, default or defined by rules, are the same in the Radix console, returned by job notifications and Job Manager API. If rules are changed, they will be applied on next deployment of an application environment, also affecting already existing batches statuses in this environment.
batchStatusRules
can be overridden for individual environments.
monitoring
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
monitoring: true
When the monitoring
field of a job-component environment config is set to true
, is used to expose custom application metrics for the specific environment.
See monitoring for more information.
monitoringConfig
spec:
jobs:
- name: compute
ports:
- name: metrics
port: 1234
monitoringConfig:
portName: metrics
path: /my-metrics
See monitoringConfig for a component for more information.
payload
spec:
jobs:
- name: compute
schedulerPort: 8000
payload:
path: /compute/args
Job specific arguments must be sent in the request body to the job-scheduler as a JSON document with an element named payload
and a value of type string.
The data type of the payload
value is string, and it can therefore contain any type of data (text, json, binary) as long as you encode it as a string, e.g. base64, when sending it to the job-scheduler, and decoding it when reading it from the mounted file inside the job container. The content of the payload is then mounted into the job container as a file named payload
in the directory specified in the payload.path
. The max size of the payload is 1MB.
In the example above, a payload sent to the job-scheduler will be mounted as file /compute/args/payload
resources
spec:
jobs:
- name: compute
resources:
requests:
memory: "256Mi"
cpu: "400m"
limits:
cpu: "600m"
See resources for more information.
secrets
spec:
jobs:
- name: compute
schedulerPort: 8000
secrets:
- DB_PASS
See secrets for a component for more information.
resources
(common)
spec:
jobs:
- name: compute
schedulerPort: 8000
resources:
requests:
memory: "6Gi"
cpu: "1000m"
limits:
cpu: "2000m"
See resources for a component for more information.
variables
(common)
spec:
jobs:
- name: compute
schedulerPort: 8000
variables:
DB_NAME: my-db
See variables for a component for more information.
timeLimitSeconds
spec:
jobs:
- name: compute
schedulerPort: 8000
timeLimitSeconds: 120
The maximum number of seconds a job can run, with a default value of 43200
seconds (12 hours). If the job's running time exceeds the limit, a SIGTERM signal is sent to allow the job to gracefully shut down with a 30 second time limit, after which it will be forcefully terminated.
backoffLimit
spec:
jobs:
- name: compute
backoffLimit: 5
Defines the number of times a job will be restarted if its container exits in error. Once the backoffLimit
has been reached the job will be marked as Failed
. The default value is 0
.
failurePolicy
spec:
jobs:
- name: compute
backoffLimit: 5
failurePolicy:
rules:
- action: FailJob
onExitCodes:
operator: In
values: [1]
- action: Ignore
onExitCodes:
operator: In
values: [143]
environmentConfig:
- environment: prod
failurePolicy:
rules:
- action: FailJob
onExitCodes:
operator: In
values: [42]
failurePolicy
defines how job container failures should be handled based on the exit code. When a job container exits with a non-zero exit code, it is evaluated against the rules
in the order they are defined. Once a rule matches the exit code, the remaining rules are ignored, and the defined action
is performed. When no rule matches the exit code, the default handling is applied.
Possible values for action
are:
FailJob
: indicates that the job should be marked asFailed
, even ifbackoffLimit
has not been reached.Ignore
: indicates that the counter towardsbackoffLimit
should not be incremented.Count
: indicates that the job should be handled the default way. The counter towardsbackoffLimit
is incremented.
failurePolicy
can be configured on the job level, or in environmentConfig
for a specific environment. Configuration in environmentConfig
will override all rules defined on the job level.
volumeMounts
spec:
jobs:
- name: compute
volumeMounts:
- name: volume-name
path: /path/in/container/to/mount/to
blobfuse2:
container: container-name
uid: 1000
The volumeMounts
field configures volume mounts within the job-component.
See volumeMounts for more information.
imageTagName
jobs:
- name: compute
image: docker.pkg.github.com/equinor/myapp/compute:{imageTagName}
imageTagName: master-latest
The imageTagName
allows for flexible configuration of fixed images, built outside of Radix. It can be configured with separate tag for each environment.
See imageTagName for a component for more information.
environmentConfig
The environmentConfig
section is to set environment-specific settings for each job.
notifications
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
notifications:
webhook: http://api:8080/monitor-batch-status
See notifications for a component for more information.
batchStatusRules
spec:
jobs:
- name: compute
batchStatusRules:
- condition: All
operator: NotIn
jobStatuses:
- Waiting
- Active
- Running
batchStatus: Completed
environmentConfig:
- environment: prod
batchStatusRules:
- condition: All
operator: In
jobStatuses:
- Succeeded
batchStatus: Succeeded
When batchStatusRules
is defined for an environment it fully overrides the job's batchStatusRules
.
See batchStatusRules for a job for more information.
monitoring
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
monitoring: true
When the monitoring
field of a job-component environment config is set to true
, is used to expose custom application metrics for the specific environment.
See monitoring for more information.
resources
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
resources:
requests:
memory: "6Gi"
cpu: "1000m"
limits:
cpu: "2000m"
See resources for a component for more information.
variables
spec:
jobs:
- name: compute
environmentConfig:
- environment: dev
variables:
DB_HOST: "db-dev"
DB_PORT: "1234"
- environment: prod
variables:
DB_HOST: "db-prod"
DB_PORT: "9876"
See variables for a component for more information.
imageTagName
jobs:
- name: compute
image: docker.pkg.github.com/equinor/myapp/compute:{imageTagName}
environmentConfig:
- environment: qa
imageTagName: master-latest
- environment: prod
imageTagName: release-39f1a082
The imageTagName
can be configured with separate tag for each environment. Environment imageTagName
overrides a job-component imageTagName
if it is also defined.
See imageTagName for a component for more information.
volumeMounts
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
volumeMounts:
- name: volume-name
path: /path/in/container/to/mount/to
blobfuse2:
container: container-name
uid: 1000
The volumeMounts
field configures volume mounts within the job-component running in the specific environment. EnvironmentConfig volumeMounts
combine or override a job-component volumeMounts
if they are defined.
See volumeMounts for more information.
timeLimitSeconds
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
timeLimitSeconds: 130
See timeLimitSeconds for more information.
backoffLimit
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
backoffLimit: 10
See backoffLimit for more information.
failurePolicy
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
backoffLimit: 5
failurePolicy:
rules:
- action: FailJob
onExitCodes:
operator: In
values: [42]
- action: Count
onExitCodes:
operator: In
values: [1, 2, 3]
- action: Ignore
onExitCodes:
operator: In
values: [143]
See failurePolicy for more information.
readOnlyFileSystem
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
readOnlyFileSystem: true|false
See readOnlyFileSystem for more information.
runtime
spec:
jobs:
- name: compute
environmentConfig:
- environment: prod
runtime:
architecture: amd64|arm64
See runtime for more information.
identity
spec:
jobs:
- name: compute
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000000
environmentConfig:
- environment: prod
identity: ...
See identity for more information.
readOnlyFileSystem
spec:
jobs:
- name: compute
readOnlyFileSystem: true|false
environmentConfig:
- environment: prod
readOnlyFileSystem: true|false
See readOnlyFileSystem for more information.
runtime
spec:
jobs:
- name: compute
runtime:
architecture: amd64|arm64
environmentConfig:
- environment: prod
runtime:
architecture: amd64|arm64
See runtime for more information.
dnsAppAlias
spec:
dnsAppAlias:
environment: prod
component: frontend
As a convenience for nicer URLs, dnsAppAlias
creates a DNS alias in the form of <app-name>.app.<cluster-dns-zone>
(<cluster-dns-zone>
depends on which Radix cluster is hosting the application) for the specified environment and component.
In the example above, if the application is host in the Platform (North Europe) cluster, the component frontend in environment prod will be accessible from myapp.app.radix.equinor.com
, in addition to automatically allocated domain names.
dnsAlias
spec:
dnsAlias:
- alias: myapp
environment: prod
component: frontend
dnsAlias
creates one or several DNS aliases in the form of <alias>.<cluster-dns-zone>
(<cluster-dns-zone>
depends on which Radix cluster is hosting the application) for the specified environment and component. There are few reserved aliases which cannot be used:
www
, app
, api
, console
, webhook
, playground
, dev
, grafana
, prometheus
, canary
, cost-api
.
In the example above, if the application is host in the Platform (North Europe) cluster, the component frontend in environment prod will be accessible from myapp.radix.equinor.com
, in addition to automatically allocated domain names.
dnsExternalAlias
spec:
dnsExternalAlias:
- alias: some.alias.com
environment: prod
component: frontend
useCertificateAutomation: [false|true]
- alias: another.alias.com
environment: prod
component: frontend
useCertificateAutomation: [false|true]
It is possible to have multiple custom DNS aliases (i.e. to choose your own custom domains) for the application. The dnsExternalAlias
needs to point to a component marked as public. It can be any domain name, which can in turn be used for public URLs to access the application.
In the example above, the component frontend hosted in environment prod will be accessible from both some.alias.com
and another.alias.com
, as long as the correct certificate has been set.
The useCertificateAutomation
property defines how the TLS certificate for the alias is issued. When set to true
, certificate issuing and renewal is automatically handled by Radix, and when set to false
, the application administrator is responsible for providing a valid certificate and private key. If the value is toggled, Radix deletes the existing certificate and private key. This will cause a TLS error when accessing the external DNS alias until a new certificate and private key is set.
There is a detailed guide on how to set up external aliases.
privateImageHubs
spec:
components:
- name: webserver
image: myappacr.azurecr.io/myapp-base:latest
privateImageHubs:
myappacr.azurecr.io:
username: 23452345-3d71-44a7-8476-50e8b281abbc
email: radix@statoilsrm.onmicrosoft.com
It is possible to pull images from private image hubs during deployment for an application. This means that you can add a reference to a private image hub in radixconfig.yaml file using the image:
tag. See example above.
With the useBuildKit: true
setting in your radixconfig.yaml
, you can also use privateImageHub credentials within the Dockerfile FROM
instruction.
FROM myappacr.azurecr.io/myapp-base:latest
A password
for these must be set via the Radix Web Console (under Configuration -> Private image hubs).
To get more information on how to connect to a private Azure container registry (ACR), see the following guide. The chapter Provisioning an Azure Container Registry
provide information on how to get service principle username
and password
. It is also possible to create a Service Principle in Azure AD, and then manually grant it access to your ACR.
See guide on how make use of privateImageHubs
in a deploy-only scenario.
node
node
section describes settings of Kubernetes node on which Radix application components or jobs are scheduled to run.
gpu
spec:
components:
- name: server
ports:
- name: http
port: 3000
node:
gpu: nvidia-v100, -nvidia-k80
gpuCount: 4
jobs:
- name: dev
node:
gpu: nvidia-k80
gpuCount: 2
When a component should run on a Kubernetes node with a GPU card on it, this need to be specified in the gpu
key of the node
section.
node:
gpu: nvidia-v100, nvidia-p100
Put one or multiple (comma separated) GPU types, which is currently supported by Radix Kubernetes cluster and which fits to component logic, which requires GPU. Currently available nodes with GPUs:
nvidia-v100
with 1, 2 or 4 GPU-s per node
node:
gpu: nvidia-v100, -nvidia-k80
When particular type of GPUs do not fit to component's logic - prefix GPU type with minus -
, component will not be scheduled on nodes with such GPU types.
node:
gpu: nvidia-v100
gpuCount: 4
When the component required multiple GPUs available on a node - put required minimum GPU count in the gpuCount
key (default value is 1
).
When gpuCount
is specified, but gpu
key is not set - component will be running on a node with any type of available GPU, which has required amount of GPUs.
node:
gpuCount: 4
secretRefs
azureKeyVault
Azure Key Vault secrets, keys and certificates can be used in Radix as secrets. Once configured, they are available in Radix application component replicas as environment variables and files content.
secretRefs:
azureKeyVaults:
- name: radix-app-secrets
path: /mnt/key-vault
useAzureIdentity: true
items:
- name: connection-string-dev
type: secret
envVar: CONNECTION_STRING
- name: key1
type: key
envVar: KEY1
- name: cert1
type: cert
envVar: CERT1
azureKeyVaults
- list of Azure Key Vault configurations.name
- Name of the Key Vault resource in an Azure subscription. Radix supports capital letters in the name, but not spaces.path
- Folder path in running replica container, where secrets, keys and/or certificate contents are available as files (with file names, corresponding to their names in the Azure Key Vault). This field is optional. If set, it overrides default path:/mnt/azure-key-vault/<azure-key-vault-name>
.useAzureIdentity
- If set totrue
, Radix will use Azure Workload Identity to acquire credentials for accessing Azure Key Vault using the service principal configured in identity.azure. This field is optional, with default valuefalse
. If omitted or set tofalse
, credentials are acquired using Azure Service Principal Client ID and Client Secret.items
- list of secrets, keys and/or certificates with corresponding environment variable names.name
- name of secret, key or certificate in an Azure Key Vault.type
- Type of the item in the Azure Key Vault. Possible values:secret
,key
,cert
. This field is optional, by default it issecret
.envVar
- Name of an environment variable, which will contain specified secret, key or certificate. This field is optional - environment variable is not created if it is not specified, only file exist (see the propertypath
).alias
- Alias of the file (see the propertypath
). This field is optional. Default value is the same asname
.
secretRefs
can be configured for entire component, for component environments or only for specific component environments. Configuration in component environments overrides similar component properties.
Updated values of secrets, keys or certificates in Azure Key Vault are not automatically synced to corresponding secrets of already running replicas - they can be synced with a new deployment.
Azure Key Vaults configurable the same way in job-components.
See guide on how to configure Azure Key Vault in Radix.
Example radixconfig.yaml
file
This example showcases all options; in many cases the defaults will be a good choice instead.
apiVersion: radix.equinor.com/v1
kind: RadixApplication
metadata:
name: myapp
spec:
build:
secrets:
- SECRET_1
- SECRET_2
environments:
- name: dev
build:
from: master
- name: prod
egress:
allowRadix: true
rules:
- destinations:
- "143.97.5.5/32"
- "143.97.6.1/32"
ports:
- port: 443
protocol: TCP
components:
- name: frontend
src: frontend
ports:
- name: http
port: 8000
- name: metrics
port: 8060
publicPort: http
monitoringConfig:
portName: metrics
path: /api/my-magic-metrics
authentication:
clientCertificate:
verification: "optional_no_ca"
node:
gpu: nvidia-v100
gpuCount: 4
enabled: true
volumeMounts:
- name: volume-name
path: /path/in/container/to/mount/to
blobfuse2:
container: container-name
uid: 1000
secretRefs:
azureKeyVaults:
- name: radix-app-secrets
path: /mnt/key-vault
useAzureIdentity: true
items:
- name: connection-string-dev
type: secret
envVar: CONNECTION_STRING
- name: key1
type: key
envVar: KEY1
- name: cert1
type: cert
envVar: CERT1
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000000
environmentConfig:
- environment: prod
monitoring: true
resources:
requests:
memory: "64Mi"
cpu: "100m"
limits:
cpu: "200m"
authentication:
clientCertificate:
passCertificateToUpstream: true
enabled: false
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000001
- environment: dev
monitoring: false
resources:
requests:
memory: "128Mi"
cpu: "200m"
limits:
cpu: "400m"
authentication:
clientCertificate:
verification: "off"
passCertificateToUpstream: false
- name: backend
src: backend
ports:
- name: http
port: 5000
environmentConfig:
- environment: dev
variables:
DB_HOST: "db-dev"
DB_PORT: "1234"
- environment: prod
replicas: 2
variables:
DB_HOST: "db-prod"
DB_PORT: "9876"
secrets:
- DB_PASS
jobs:
- name: compute
schedulerPort: 8000
ports:
- name: http
port: 9000
payload:
path: /compute/args
timeLimitSeconds: 300
backoffLimit: 2
resources:
requests:
memory: "256Mi"
cpu: "400m"
limits:
cpu: "600m"
variables:
DB_NAME: "compute-db"
secrets:
- DB_USER
- DB_PASS
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000000
notifications:
webhook: http://backend:5000/monitor-batch-status
environmentConfig:
- environment: dev
variables:
DB_HOST: "db-dev"
DB_PORT: "1234"
notifications:
webhook: ""
- environment: prod
timeLimitSeconds: 600
backoffLimit: 10
monitoring: true
variables:
DB_HOST: "db-prod"
DB_PORT: "1234"
identity:
azure:
clientId: 00000000-0000-0000-0000-000000000001
notifications:
webhook: http://backend:5000/monitor-batch-status-prod
dnsAppAlias:
environment: prod
component: frontend
dnsExternalAlias:
- alias: some.alias.com
environment: prod
component: frontend
- alias: another.alias.com
environment: prod
component: frontend