Updated 15/02/2024
DTE Infrastructure Component

interLink

Federated Compute
Image

Description

An open-source service to enable transparent access to heterogeneous computing providers

InterLink is an open-source service to enable transparent access to heterogeneous computing providers. It provides an abstraction for the execution of a Kubernetes pod on any remote resource capable of managing a Container execution lifecycle. The interLink component extends the Kubernetes Virtual Kubelet solution with a generic API layer for delegating pod execution on ANY remote backend. Kubernetes POD requests are digested through the API layer (e.g. deployed on an HPC edge) into batch job execution of a container.

The API layer foresees a plugin structure to accommodate any possible backend integration.

Executing payloads in response to an external trigger like a storage event or a web server call.

Frameworks for DAG workflow managements are usually well integrated with Kubernetes APIs

Users can exploit interLink via self-provisioned deployment (i.e. through already integrated high level services) or standalone Kubernetes deployment creating and deploying a simple container that can be scheduled on a remote system such as a Slurm batch on an HPC center. High Level services getting integrated are e.g.: Airflow/Kubeflow pipelines/Argo workflows/MLFlow, Jupyter notebooks.

Target Audience
+

Scientific users adopting Docker container-based workflow management or pipeline management system.

License
+

Apache 2.0

Created by
+

Release Notes

interLink is an open-source software product under development in the context of interTwin project. Its development started in order to provide an open-source solution capable of extending the container orchestration de-facto standard (kubernetes) to support offloading to any type of resource provider (Cloud/HTC/HPC) in a transparent manner, where little to no knowledge is required for the end user. The key objective of interLink is to enable a Kubernetes cluster to send containers/pod to a “virtual” node. This node seamlessly manages the entire lifecycle of the user’s applications, whether on a remote server or, preferably, within an HPC batch queue.

Transparency is granted by the fact that it keeps exposing the very same experience of running a pod on cloud resources, thanks to the API layer that exposes the regular set of Kubernetes APIs.

The interLink project is organised in 4 major subcomponents. Namely:

  • A Virtual Kubelet
  • The InterLink API
  • Plugins
  • Monitoring system components

Future Plans

Future Plans

Feature wise the system is almost complete. The main priority for the upcoming period  is on further integrating the support for the Datalake and data management interoperability.

The current logging capabilities are providing a wide set of information about the status of the Virtual Kubelets and this allows easier debugging in case of problems. However this is not enough and thus on the one side we plan to extend the management of the collected info while on the other side we plan to implement a full set of monitoring handles on the interLink API layer. The latter also includes the Plugin structure.

Additional improvements currently planned are related to the integration with ALISE service. These updates aim to address site resources access policies and improve the management of File System permissions on remote resources.

Furthermore during the last period of the project the plan is to implement extensive scale tests. This will be a key process and through all this we expect to bring the system to TRL 6.

 

Related Use Cases

Image
interTwin Use Case: A Digital Twin for Drought Early Warning in the Alps

Developing the components to set up a Drought Early Warning System for the Alps

...
Image
interTwin Use Case: A Digital Twin for Lattice QCD simulation

Applying Machine Learning techniques to Lattice Simulation to speed up data generation and open up...

Image
interTwin Use Case: A Digital Twin for projecting the occurrence of tropical cyclones due to climate change

Designing Deep neural networks (DNNs) to create Digital Twins capable of detecting tropical cyclones...

Image
interTwin Use Case: A Digital Twin to simulate ‘noise’ in the Virgo Gravitational Wave interferometer

Creating a Digital Twin to realistically simulate ‘noise’ in the detector and study how it...

Image
interTwin Use Case: A Digital Twin for Flood Early Warning in coastal and inland regions

Developing the components to set up a Digital Twins for flood early warning in coastal...

Image
interTwin Use Case: A Digital Twin for projecting wildfire danger due to climate change

Defining deep neural networks (DNNs) for creating Digital Twins for predicting fire danger on climate...

Image
interTwin Use Case: A Digital Twin to simulate ‘noise’ in Radio Astronomy

Building a digital twin of a source-telescope system that provides synthetic data, which can be...

Image
interTwin Use Case: A particle detector data-driven Digital Twin for High Energy Physics

Setting up a Digital Twin for fast particle detector simulation

...
Image
interTwin Use Case: Deploying FloodAdapt, a Digital Twin for flood impact modelling, anywhere on earth

Developing the necessary components to facilitate the automatic set up of a Digital Twin supporting...

Image
interTwin Use Case: Extreme rainfall, temperature and wind – weather extremes as a result of climate change

Using AI to create a Digital Twin Application that can detect and characterize the changes...

Image