Serverless computing continues to be a much-discussed technology, in large part due to the perception that it enables organizations to focus more clearly on their applications, instead of the underlying infrastructure. More specifically, serverless “refers to the concept of building and running applications that do not require server management. It describes a finer-grained deployment model where applications and microservices, bundled as functions or not, are uploaded to a platform and then executed, scaled, and billed in response to the exact demand needed at the moment,” as defined by the Cloud Native Computing Foundation (CNCF) community.
Today marks a milestone for serverless, specifically Knative. While Knative serverless primitives have been available since July 2018 to give users a more portable way to write serverless applications, today marks a commitment to making the technology enterprise-ready. Technology leaders including Red Hat, Google, SAP and IBM, are committing to leveraging Knative as an open standard in their commercial serverless offerings. This common set of constructs in Knative enables interoperability for serverless workloads on Kubernetes installations. Coupled with Red Hat OpenShift, Knative can further enable portability of operations in hybrid environments.
Red Hat OpenShift - our industry-leading enterprise Kubernetes platform - plans to add support for Knative, in dev preview early next year. This is designed to enable users to build and run serverless applications, integrating with Red Hat OpenShift Service Mesh, based on the Istio and Kiali projects. We are also using Strimzi, which makes it easier to run Apache Kafka on OpenShift or Kubernetes, via Red Hat AMQ Streams for reliable eventing and Camel-K, a lightweight integration framework built from Apache Camel that enables multiple event sources to be used as triggers for serverless applications. Today, users will be able to get started with Knative on minishift by reproducing our demo.
Looking ahead, Red Hat plans to take this a step further to enable consistent operations across any cloud provider for Knative-based serverless applications spanning the ecosystem of ISVs with Certified OpenShift Services.
Google Cloud introduced Knative, an open source project that builds on top of the common platform capabilities of Kubernetes, including monitoring, logging, identity management, and security, earlier this year. Knative provides a set of middleware components for building modern, source-centric, container-based applications that can run on-premises or in the cloud. The model enables developers to write and manage serverless applications in a consistent way and is both developer and operator-friendly. Red Hat is among the major vendors who have been using its expertise in enterprise Kubernetes into those building blocks making sure that they are well suited to developer needs. Knative is also compatible with other emerging technologies, such as Istio.
Red Hat is committed to providing support for hybrid application development and operations and that now also includes serverless workloads. Using the Operator Lifecycle Manager and the Operator Framework, the way you deploy, manage and upgrade serverless platforms, event sources and applications can be consistent across any cloud provider, as well as on-premises. This not only provides consistency but it also enables interoperability.
At this week’s KubeCon, we are discussing how Red Hat is working to bring Knative CRDs to developers, enabling ISVs and their services to be consumed by applications and functions that are written in an opinionated model. Simply put, this model can enable event-driven applications to inherit serverless traits, including scale-from-zero, on-demand events following industry best practices for modern, cloud-native application development. This keeps code and configuration separate, and allows for disposable resources, provisioning and a pluggable build model. For OpenShift, it means that developers can reuse toolings that they are already familiar with such as OpenShift builds, source-to-image (S2I) and Buildah.
Along with Google Cloud and other key enterprise players, Red Hat intends to take the interoperability of serverless workloads to the next level. Between OpenShift and Knative, other containerized applications can benefit from the serverless, hybrid cloud.
“Introduced in July, Knative is an open source project based on Kubernetes that provides critical building blocks for modern serverless applications,” said Oren Teich, Director of Product Management, Google Cloud. “In the four months since then we have seen tremendous enthusiasm from contributing companies, many of whom are furthering Knative as the serverless standard for hybrid and multi-cloud users. Red Hat has been a contributor to Knative from the start, and the integration of Knative in Openshift is a key outcome of our joint open source efforts.”
For more info, see Google’s blog. For more on our work with Knative and serverless, please see the first post in our blog series on Red Hat OpenShift, Knative: Serving your Serverless Services.
Über den Autor
William Markito Oliveira is an energetic and passionate product leader with expertise in software engineering and distributed systems. He leads a group of product managers working on innovative and emerging technologies.
He has worked in global and distributed organizations, with success building high-performance teams and successful products. He also has experience establishing internal and external partnerships delivering high-value outcomes. Oliveira is engaged in open source by fostering communities and governance, and he's looking forward to embracing new technologies and paradigms with a focus on distributed systems. He's spoken at multiple software conferences and co-authored a few books on the field.
Currently, Oliveira is focused on the intersection between serverless, cloud computing, and Kubernetes and applying AI/ML concepts whenever possible.
Nach Thema durchsuchen
Automatisierung
Das Neueste zum Thema IT-Automatisierung für Technologien, Teams und Umgebungen
Künstliche Intelligenz
Erfahren Sie das Neueste von den Plattformen, die es Kunden ermöglichen, KI-Workloads beliebig auszuführen
Open Hybrid Cloud
Erfahren Sie, wie wir eine flexiblere Zukunft mit Hybrid Clouds schaffen.
Sicherheit
Erfahren Sie, wie wir Risiken in verschiedenen Umgebungen und Technologien reduzieren
Edge Computing
Erfahren Sie das Neueste von den Plattformen, die die Operations am Edge vereinfachen
Infrastruktur
Erfahren Sie das Neueste von der weltweit führenden Linux-Plattform für Unternehmen
Anwendungen
Entdecken Sie unsere Lösungen für komplexe Herausforderungen bei Anwendungen
Original Shows
Interessantes von den Experten, die die Technologien in Unternehmen mitgestalten
Produkte
- Red Hat Enterprise Linux
- Red Hat OpenShift
- Red Hat Ansible Automation Platform
- Cloud-Services
- Alle Produkte anzeigen
Tools
- Training & Zertifizierung
- Eigenes Konto
- Kundensupport
- Für Entwickler
- Partner finden
- Red Hat Ecosystem Catalog
- Mehrwert von Red Hat berechnen
- Dokumentation
Testen, kaufen und verkaufen
Kommunizieren
Über Red Hat
Als weltweit größter Anbieter von Open-Source-Software-Lösungen für Unternehmen stellen wir Linux-, Cloud-, Container- und Kubernetes-Technologien bereit. Wir bieten robuste Lösungen, die es Unternehmen erleichtern, plattform- und umgebungsübergreifend zu arbeiten – vom Rechenzentrum bis zum Netzwerkrand.
Wählen Sie eine Sprache
Red Hat legal and privacy links
- Über Red Hat
- Jobs bei Red Hat
- Veranstaltungen
- Standorte
- Red Hat kontaktieren
- Red Hat Blog
- Diversität, Gleichberechtigung und Inklusion
- Cool Stuff Store
- Red Hat Summit