# intent-driven-orchestration **Repository Path**: mirrors_intel/intent-driven-orchestration ## Basic Information - **Project Name**: intent-driven-orchestration - **Description**: Intent Driven Orchestration enables management of applications through their Service Level Objectives, while minimizing developer and administrator overhead. - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2022-10-24 - **Last Updated**: 2026-05-02 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Intent Driven Orchestration Planner ![planner.png](planner.png) Today’s container orchestration engine solutions promote a model of requesting a specific quantity of resources (e.g. number of vCPUs.), a quantity range (e.g. min/max number of vCPUs) or not specifying them at all, to support the appropriate placement of workloads. This applies at the Cloud and at the Edge using [Kubernetes](https://kubernetes.io/) or [K3s](https://k3s.io/) (although the concept is not limited to Kubernetes based systems). The end state for the resource allocation is declared, but that state is an imperative definition of what resources are required. This model has proven effective, but has a number of challenges: * experience shows that often suboptimal information is declared leading to over allocating resource or sub-optimal performance, * these declarations have context that the users needs to understand, but often do not (e.g. resource requests do not dynamically change based on the chosen instance type that underpins a cluster). This project proposes a new way to do orchestration – moving from an imperative model to an intent driven way, in which the user express their intents in form of objectives (e.g. as required latency, throughput, or reliability targets) and the orchestration stack itself determines what resources in the infrastructure are required to fulfill the objectives. This new approach will continue to benefit from community investments in _**scheduling**_ (determining **when & where** to place workloads) and be augmented with a continuous running **_planning_** loop determining **what/how** to configure in the system. While this repository holds the planning component implementation it is key to note that it works closely together with schedulers, the observability and potentially analytics stacks. It is key that those schedulers are fed with the right information to make their placement decisions. The planning component is essential for enabling Intent Driven Orchestration (IDO), as it will break down the higher- level objectives (e.g. a latency compliance targets) into dynamic actionable plans (e.g. policies for platform resource allocation, dynamic vertical & horizontal scaling, etc.). This enables hierarchical controlled systems in which Service Level Objectives(SLOs) are broken down to finer grained goal settings for the platform. A key input to the planning components, to determine the right set of actions, are the models that describe workload behaviour and the platform effects on the associated Quality of Service (QoS). The initial goal is to focus on managing the QoS of a set of instances of a workload. Subsequently, the goals are expected to shift to End-to-End (E2E) management of QoS parameters in multi-tenant environments with mixed criticality. It is also a goal that the planning components will be easily extended and administrators will have the ability to swap in and out functionality through a plugin model. The architecture is intended to be extensible to support proactive planning and coordination between planners to fulfill overarching intents. It is expected that the imperative model and an Intent Driven Orchestration model will coexist. ## Example To see the benefit of this model please review the deployment and associated objective manifest files: * The Deployment [spec](artefacts/examples/example_deployment.yaml) only defines which container image to use (hence fully abstracted from platform resources. This is different but still analogous to Serverless methodology). * The Objective [spec](artefacts/examples/example_intent.yaml) defines a P-95 latency compliance target of less than 4ms as measured by e.g. a Service Mesh and requests an availability target of 2 nines (i.e. 99% availability). ## Running the planning component Step 1) add the CRDs: $ k apply -f artefacts/intents_crds_v1alpha1.yaml Step 2) deploy the planner (make sure to adapt the configs to your environment): $ k create ns ido $ k apply -n ido -f artefacts/deploy/manifest.yaml Step 3) deploy the actuators of interest using: $ k apply -n ido -f plugins//.yaml These steps should be followed by setting up your default profiles (if needed). We recommend the usage of a service mesh like [Linkerd](https://linkerd.io/) or [Istio](https://istio.io/) to ensure robust authentication, encryption and monitoring capabilities for the subcomponents of the planning framework themselves. After creating the namespace,