Duration 2 Days 12 CPD hours This course is intended for This class is primarily intended for the following participants: Technical employees using GCP, including customer companies, partners and system integrators deployment engineers, cloud architects, cloud administrators, system engineers , and SysOps/DevOps engineers. Individuals using GCP to create, integrate, or modernize solutions using secure, scalable microservices architectures in hybrid environments. Overview Connect and manage Anthos GKE clusters from GCP Console whether clusters are part of Anthos on Google Cloud or Anthos deployed on VMware. Understand how service mesh proxies are installed, configured and managed. Configure centralized logging, monitoring, tracing, and service visualizations wherever the Anthos GKE clusters are hosted. Understand and configure fine-grained traffic management. Use service mesh security features for service-service authentication, user authentication, and policy-based service authorization. Install a multi-service application spanning multiple clusters in a hybrid environment. Understand how services communicate across clusters. Migrate services between clusters. Install Anthos Config Management, use it to enforce policies, and explain how it can be used across multiple clusters. This two-day instructor-led course prepares students to modernize, manage, and observe their applications using Kubernetes whether the application is deployed on-premises or on Google Cloud Platform (GCP). Through presentations, and hands-on labs, participants explore and deploy using Kubernetes Engine (GKE), GKE Connect, Istio service mesh and Anthos Config Management capabilities that enable operators to work with modern applications even when split among multiple clusters hosted by multiple providers, or on-premises. Anthos Overview Describe challenges of hybrid cloud Discuss modern solutions Describe the Anthos Technology Stack Managing Hybrid Clusters using Kubernetes Engine Understand Anthos GKE hybrid environments, with Admin and User clusters Register and authenticate remote Anthos GKE clusters in GKE Hub View and manage registered clusters, in cloud and on-premises, using GKE Hub View workloads in all clusters from GKE Hub Lab: Managing Hybrid Clusters using Kubernetes Engine Introduction to Service Mesh Understand service mesh, and problems it solves Understand Istio architecture and components Explain Istio on GKE add on and it's lifecycle, vs OSS Istio Understand request network traffic flow in a service mesh Create a GKE cluster, with a service mesh Configure a multi-service application with service mesh Enable external access using an ingress gateway Explain the multi-service example applications: Hipster Shop, and Bookinfo Lab: Installing Open Source Istio on Kubernetes Engine Lab: Installing the Istio on GKE Add-On with Kubernetes Engine Observing Services using Service Mesh Adapters Understand service mesh flexible adapter model Understand service mesh telemetry processing Explain Stackdriver configurations for logging and monitoring Compare telemetry defaults for cloud and on-premises environments Configure and view custom metrics using service mesh View cluster and service metrics with pre-configured dashboards Trace microservice calls with timing data using service mesh adapters Visualize and discover service attributes with service mesh Lab: Telemetry and Observability with Istio Managing Traffic Routing with Service Mesh Understand the service mesh abstract model for traffic management Understand service mesh service discovery and load balancing Review and compare traffic management use cases and configurations Understand ingress configuration using service mesh Visualize traffic routing with live generated requests Configure a service mesh gateway to allow access to services from outside the mesh Apply virtual services and destination rules for version-specific routing Route traffic based on application-layer configuration Shift traffic from one service version to another, with fine-grained control, like a canary deployment Lab: Managing Traffic Routing with Istio and Envoy Managing Policies and Security with Service Mesh Understand authentication and authorization in service mesh Explain mTLS flow for service to service communication Adopt mutual TLS authentication across the service mesh incrementally Enable end-user authentication for the frontend service Use service mesh access control policies to secure access to the frontend service Lab: Managing Policies and Security with Service Mesh Managing Policies using Anthos Config Management Understand the challenge of managing resources across multiple clusters Understand how a Git repository is as a configuration source of truth Explain the Anthos Config Management components, and object lifecycle Install and configure Anthos Config Management, operators, tools, and related Git repository Verify cluster configuration compliance and drift management Update workload configuration using repo changes Lab: Managing Policies in Kubernetes Engine using Anthos Config Configuring Anthos GKE for Multi-Cluster Operation Understand how multiple clusters work together using DNS, root CA, and service discovery Explain service mesh control-plane architectures for multi-cluster Configure a multi-service application using service mesh across multiple clusters with multiple control-planes Configure a multi-service application using service mesh across multiple clusters with a shared control-plane Configure service naming/discovery between clusters Review ServiceEntries for cross-cluster service discovery Migrate workload from a remote cluster to an Anthos GKE cluster Lab: Configuring GKE for Multi-Cluster Operation with Istio Lab: Configuring GKE for Shared Control Plane Multi-Cluster Operation
Duration 5 Days 30 CPD hours This course is intended for Audience for this course This course is designed for system administrators responsible for creating OpenShift Enterprise instances, deploying applications, creating process customizations, managing instances and projects. Prerequisites for this course Have taken Red Hat Enterprise Linux Administration I and II (RH124 and RH134), or equivalent Red Hat Enterprise Linux© system administration experience Be certified as a Red Hat Certified System Administrator (RHCSA), or equivalent Red Hat Enterprise Linux system administration experience Be certified as a Red Hat Certified Engineer (RHCE©) Overview Learn to install, configure, and manage OpenShift Enterprise by Red Hat instances - OpenShift Enterprise Administration (DO280) prepares the system administrator to install, configure, and manage OpenShift Enterprise by Red Hat© instances. OpenShift Enterprise, Red Hat's platform-as-a-service (PaaS) offering, provides pre-defined deployment environments for applications of all types through its use of container technology. This creates an environment that supports DevOps principles such as reduced time to market and continuous delivery. - In this course, students will learn how to install and configure an instance of OpenShift Enterprise, test the instance by deploying a real world application, and manage projects/applications through hands-on labs. - Course content summary - Container concepts - Configuring resources with the command line interface - Building a pod - Enabling services for a pod - Creating routes - Downloading and configuring images - Rolling back and activating deployments - Creating custom S2I images This course will empower you to install and administer the Red Hat© OpenShift© Container Platform, with hands-on, lab-based materials that show you how to install, configure, and manage OpenShift clusters and deploy sample applications to further understand how developers will use the platform. This course is based on Red Hat© Enterprise Linux© 7.5 and Openshift Container Platform 3.9. OpenShift is a containerized application platform that allows your enterprise to manage container deployments and scale your applications using Kubernetes. OpenShift provides predefined application environments and builds upon Kubernetes to provide support for DevOps principles such as reduced time to market, infrastructure-as-code, continuous integration (CI), and continuous delivery (CD). 1 - INTRODUCTION TO RED HAT OPENSHIFT ENTERPRISE Review features and architecture of OpenShift Enterprise. 2 - INSTALL OPENSHIFT ENTERPRISE Install OpenShift Enterprise and configure a master and node. 3 - EXECUTE COMMANDS Execute commands using the command line interface. 4 - BUILD APPLICATIONS Create, build, and deploy applications to an OpenShift Enterprise instance. 5 - PERSISTENT STORAGE Provision persistent storage and use it for the internal registry. 6 - BUILD APPLICATIONS WITH SOURCE-TO-IMAGE (S2I) Create and build applications with S2I and templates. 7 - MANAGE THE SYSTEM Use OpenShift Enterprise components to manage deployed applications. 8 - CUSTOMIZE OPENSHIFT ENTERPRISE Customize resources and processes used by OpenShift Enterprise. 9 - COMPREHENSIVE REVIEW Practice and demonstrate knowledge and skills learned in the course. 10 - NOTE: Course outline is subject to change with technology advances and as the nature of the underlying job evolves. For questions or confirmation on a specific objective or topic, please contact us. Additional course details: Nexus Humans Red Hat OpenShift Administration II: Operating a Production Kubernetes Cluster (DO280) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Red Hat OpenShift Administration II: Operating a Production Kubernetes Cluster (DO280) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for the following participants: Application developers, Cloud Solutions Architects, DevOps Engineers, IT managers. Individuals using Google Cloud Platform to create new solutions or to integrate existing systems, application environments, and infrastructure with the Google Cloud Platform. Overview At the end of the course, students will be able to: Understand container basics. Containerize an existing application. Understand Kubernetes concepts and principles. Deploy applications to Kubernetes using the CLI. Set up a continuous delivery pipeline using Jenkins Learn to containerize workloads in Docker containers, deploy them to Kubernetes clusters provided by Google Kubernetes Engine, and scale those workloads to handle increased traffic. Students will also learn how to continuously deploy new code in a Kubernetes cluster to provide application updates. Introduction to Containers and Docker Acquaint yourself with containers, Docker, and the Google Container Registry. Create a container. Package a container using Docker. Store a container image in Google Container Registry. Launch a Docker container. Kubernetes Basics Deploy an application with microservices in a Kubernetes cluster. Provision a complete Kubernetes cluster using Kubernetes Engine. Deploy and manage Docker containers using kubectl. Break an application into microservices using Kubernetes? Deployments and Services. Deploying to Kubernetes Create and manage Kubernetes deployments. Create a Kubernetes deployment. Trigger, pause, resume, and rollback updates. Understand and build canary deployments. Continuous Deployment with Jenkins Build a continuous delivery pipeline. Provision Jenkins in your Kubernetes cluster. Create a Jenkins pipeline. Implement a canary deployment using Jenkins. Additional course details: Nexus Humans Getting Started with Google Kubernetes Engine training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Getting Started with Google Kubernetes Engine course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 3 Days 18 CPD hours This course is intended for This class is intended for the following participants: Cloud architects, administrators, and SysOps/DevOps personnel Individuals using Google Cloud Platform to create new solutions or to integrate existing systems, application environments, and infrastructure with the Google Cloud Platform. Overview This course teaches participants the following skills: Understand how software containers work Understand the architecture of Kubernetes Understand the architecture of Google Cloud Platform Understand how pod networking works in Kubernetes Engine Create and manage Kubernetes Engine clusters using the GCP Console and gcloud/ kubectl commands Launch, roll back and expose jobs in Kubernetes Manage access control using Kubernetes RBAC and Google Cloud IAM Managing pod security policies and network policies Using Secrets and ConfigMaps to isolate security credentials and configuration artifacts Understand GCP choices for managed storage services Monitor applications running in Kubernetes Engine This class introduces participants to deploying and managing containerized applications on Google Kubernetes Engine (GKE) and the other services provided by Google Cloud Platform. Through a combination of presentations, demos, and hands-on labs, participants explore and deploy solution elements, including infrastructure components such as pods, containers, deployments, and services; as well as networks and application services. This course also covers deploying practical solutions including security and access management, resource management, and resource monitoring. Introduction to Google Cloud Platform Use the Google Cloud Platform Console Use Cloud Shell Define cloud computing Identify GCPs compute services Understand regions and zones Understand the cloud resource hierarchy Administer your GCP resources Containers and Kubernetes in GCP Create a container using Cloud Build Store a container in Container Registry Understand the relationship between Kubernetes and Google Kubernetes Engine (GKE) Understand how to choose among GCP compute platforms Kubernetes Architecture Understand the architecture of Kubernetes: pods, namespaces Understand the control-plane components of Kubernetes Create container images using Google Cloud Build Store container images in Google Container Registry Create a Kubernetes Engine cluster Kubernetes Operations Work with the kubectl command Inspect the cluster and Pods View a Pods console output Sign in to a Pod interactively Deployments, Jobs, and Scaling Create and use Deployments Create and run Jobs and CronJobs Scale clusters manually and automatically Configure Node and Pod affinity Get software into your cluster with Helm charts and Kubernetes Marketplace GKE Networking Create Services to expose applications that are running within Pods Use load balancers to expose Services to external clients Create Ingress resources for HTTP(S) load balancing Leverage container-native load balancing to improve Pod load balancing Define Kubernetes network policies to allow and block traffic to pods Persistent Data and Storage Use Secrets to isolate security credentials Use ConfigMaps to isolate configuration artifacts Push out and roll back updates to Secrets and ConfigMaps Configure Persistent Storage Volumes for Kubernetes Pods Use StatefulSets to ensure that claims on persistent storage volumes persist across restarts Access Control and Security in Kubernetes and Kubernetes Engine Understand Kubernetes authentication and authorization Define Kubernetes RBAC roles and role bindings for accessing resources in namespaces Define Kubernetes RBAC cluster roles and cluster role bindings for accessing cluster-scoped resources Define Kubernetes pod security policies Understand the structure of GCP IAM Define IAM roles and policies for Kubernetes Engine cluster administration Logging and Monitoring Use Stackdriver to monitor and manage availability and performance Locate and inspect Kubernetes logs Create probes for wellness checks on live applications Using GCP Managed Storage Services from Kubernetes Applications Understand pros and cons for using a managed storage service versus self-managed containerized storage Enable applications running in GKE to access GCP storage services Understand use cases for Cloud Storage, Cloud SQL, Cloud Spanner, Cloud Bigtable, Cloud Firestore, and Bigquery from within a Kubernetes application
Duration 4 Days 24 CPD hours Overview Topics Include:Installation of a multi-node Kubernetes cluster using kubeadm, and how to grow a cluster.Choosing and implementing cluster networking.Various methods of application lifecycle management, including scaling, updates and roll-backs.Configuring security both for the cluster as well as containers.Managing storage available to containers.Learn monitoring, logging and troubleshooting of containers and the cluster.Configure scheduling and affinity of container deployments.Use Helm and Charts to automate application deployment.Understand Federation for fault-tolerance and higher availability. In this vendor agnostic course, you'll learn the installation, configuration and administration of a production-grade Kubernetes cluster. Introduction Linux Foundation Linux Foundation Training Linux Foundation Certifications Laboratory Exercises, Solutions and Resources Distribution Details Labs Basics of Kubernetes Define Kubernetes Cluster Structure Adoption Project Governance and CNCF Labs Installation and Configuration Getting Started With Kubernetes Minikube kubeadm More Installation Tools Labs Kubernetes Architecture Kubernetes Architecture Networking Other Cluster Systems Labs APIs and Access API Access Annotations Working with A Simple Pod kubectl and API Swagger and OpenAPI Labs API Objects API Objects The v1 Group API Resources RBAC APIs Labs Managing State With Deployments Deployment Overview Managing Deployment States Deployments and Replica Sets DaemonSets Labels Labs Services Overview Accessing Services DNS Labs Volumes and Data Volumes Overview Volumes Persistent Volumes Passing Data To Pods ConfigMaps Labs Ingress Overview Ingress Controller Ingress Rules Labs Scheduling Overview Scheduler Settings Policies Affinity Rules Taints and Tolerations Labs Logging and Troubleshooting Overview Troubleshooting Flow Basic Start Sequence Monitoring Logging Troubleshooting Resources Labs Custom Resource Definition Overview Custom Resource Definitions Aggregated APIs Labs Kubernetes Federation Overview Federated Resources Labs Helm Overview Helm Using Helm Labs Security Overview Accessing the API Authentication and Authorization Admission Controller Pod Policies Network Policies Labs
Duration 4 Days 24 CPD hours Overview By the end of the course, you should be able to meet the following objectives: Describe how Tanzu Kubernetes Grid fits in the VMware Tanzu portfolio Describe the Tanzu Kubernetes Grid architecture Deploy and manage Tanzu Kubernetes Grid management and supervisor clusters Deploy and manage Tanzu Kubernetes Grid workload clusters Deploy, configure, and manage Tanzu Kubernetes Grid packages Perform basic troubleshooting During this four-day course, you focus on installing VMware Tanzu© Kubernetes Grid? in a VMware vSphere© environment and provisioning and managing Tanzu Kubernetes Grid clusters. The course covers how to install Tanzu Kubernetes Grid packages for image registry, authentication, logging, ingress, multipod network interfaces, service discovery, and monitoring. The concepts learned in this course are transferable for users who must install Tanzu Kubernetes Grid on other supported clouds. Course Introduction Introductions and course logistics Course objectives Introducing VMware Tanzu Kubernetes Grid Identify the VMware Tanzu products responsible for Kubernetes life cycle management and describe the main differences between them Explain the core concepts of Tanzu Kubernetes Grid, including bootstrap, Tanzu Kubernetes Grid management, supervisor, and workload clusters List the components of a Tanzu Kubernetes Grid instance VMware Tanzu Kubernetes Grid CLI and API Illustrate how to use the Tanzu CLI Define the Carvel Tool set Define Cluster API Identify the infrastructure providers List the Cluster API controllers Identify the Cluster API custom resource definitions Authentication Explain how Kubernetes manages authentication with Management clusters Explain how Kubernetes manages authentication with supervisor clusters Define Pinniped Define Dex Describe the Pinniped authentication workflow Load Balancers Illustrate how load balancing works for the Kubernetes control plane Illustrate how load balancing works for application workload Explain how Tanzu Kubernetes Grid integrates with VMware NSX Advanced Load Balancer List load balancing options available on public clouds VMware Tanzu Kubernetes Grid on vSphere List the requirements for deploying a supervisor cluster List the steps to install a Tanzu Kubernetes Grid supervisor cluster Summarize the events of a supervisor cluster creation List the requirements for deploying a management cluster List the steps to install a Tanzu Kubernetes Grid management cluster Summarize the events of a management cluster creation Demonstrate how to use commands when working with management clusters VMware Tanzu Kubernetes Grid on Public Clouds List the requirements for deploying a management cluster on AWS and Microsoft Azure List the configuration options to install a Tanzu Kubernetes Grid a management cluster on AWS and Azure Tanzu Kubernetes Workload Clusters List the steps to build a custom image Describe the available customizations Identify the options for deploying Tanzu Kubernetes Grid clusters Explain the difference between the v1alpha3 and v1beta1 APIs Explain how Tanzu Kubernetes Grid clusters are created Discuss which VMs compose a Tanzu Kubernetes Grid cluster List the pods that run on a Tanzu Kubernetes Grid cluster Describe the Tanzu Kubernetes Grid core add-ons that are installed on a cluster Tanzu Kubernetes Grid Packages Define the Tanzu Kubernetes Grid packages Explain the difference between Auto-Managed and CLI-Managed packages Define packages repositories Configuring and Managing Tanzu Kubernetes Grid Operation and Analytics Packages Describe Cert-Manager Describe the Harbor Image Registry Describe Fluent Bit Identify the logs that Fluent Bit collects Explain basic Fluent Bit configuration Describe Prometheus and Grafana Configuring and Managing Tanzu Kubernetes Grid Networking Packages Describe the Contour ingress controller Demonstrate how to install Contour on a Tanzu Kubernetes Grid cluster Describe ExternalDNS Demonstrate how to install Service Discovery with ExternalDNS Describe Multus CNI Tanzu Kubernetes Grid Day 2 Operations List the load balancer configuration options in vSphere to load balance applications Demonstrate how to configure Ingress with the NodePortLocal Mode Explain how to install VMware Tanzu Application Platform Describe life cycle management in Tanzu Kubernetes Grid Explain how backup and restore are implemented in Tanzu Kubernetes Grid Describe Velero and Restic List the steps to back up a Workload cluster using Velero and Restic Troubleshooting Tanzu Kubernetes Grid Discuss the various Tanzu Kubernetes Grid logs Identify the location of Tanzu Kubernetes Grid logs Explain the purpose of crash diagnostics Demonstrate how to check the health of a Tanzu Kubernetes Grid cluster Explain packages cleanup procedures Explain management recovery procedures Additional course details:Notes Delivery by TDSynex, Exit Certified and New Horizons an VMware Authorised Training Centre (VATC) Nexus Humans VMware Tanzu Kubernetes Grid: Install, Configure, Manage [V2.0] training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the VMware Tanzu Kubernetes Grid: Install, Configure, Manage [V2.0] course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 4 Days 24 CPD hours This course is intended for Anyone who is preparing to build and run Kubernetes clusters Overview By the end of the course, you should be able to meet the following objectives: Build, test, and publish Docker container images Become familiar with YAML files that define Kubernetes objects Understand Kubernetes core user-facing concepts, including pods, services, and deployments Use kubectl, the Kubernetes CLI, and become familiar with its commands and options Understand the architecture of Kubernetes (Control plane and its components, worker nodes, and kubelet) Learn how to troubleshoot issues with deployments on Kubernetes Apply resource requests, limits, and probes to deployments Manage dynamic application configuration using ConfigMaps and Secrets Deploy other workloads, including DaemonSets, Jobs, and CronJobs Learn about user-facing security using SecurityContext, RBAC, and NetworkPolicies This four-day course is the first step in learning about Containers and Kubernetes Fundamentals and Cluster Operations. Through a series of lectures and lab exercises, the fundamental concepts of containers and Kubernetes are presented and put to practice by containerizing and deploying a two-tier application into Kubernetes. Course Introduction Introductions and objectives Containers What and Why containers Building images Running containers Registry and image management Kubernetes Overview Kubernetes project Plugin interfaces Building Kubernetes Kubectl CLI Beyond Kubernetes Basics Kubernetes objects YAML Pods, replicas, and deployments Services Deployment management Rolling updates Controlling deployments Pod and container configurations Kubernetes Networking Networking within a pod Pod-to-Pod Networking Services to Pods ClusterIP, NodePort, and LoadBalancer Ingress controllers Service Discovery via DNS Stateful Applications in Kubernetes Stateless versus Stateful Volumes Persistent volumes claims StorageClasses StatefulSets Additional Kubernetes Considerations Dynamic configuration ConfigMaps Secrets Jobs, CronJobs Security Network policy Applying a NetworkPolicy SecurityContext runAsUser/Group Service accounts Role-based access control Logging and Monitoring Logging for various objects Sidecar logging Node logging Audit logging Monitoring architecture Monitoring solutions Octant VMware vRealize Operations Manager Cluster Operations Onboarding new applications Backups Upgrading Drain and cordon commands Impact of an upgrade to running applications Troubleshooting commands VMware Tanzu portfolio overview
Duration 5 Days 30 CPD hours This course is intended for Security Professionals working with Kubernetes Clusters Container Orchestration Engineers DevOps Professionals Overview In this course, students will learn and practice essential Kubernetes concepts and tasks in the following sections: Cloud Security Fundamentals Cluster Hardening System Hardening Minimize Microservice Vulnerabilities Supply Chain Security Disaster Recovery Secure Back-up and Restore This class prepares students for the Certified Kubernetes Security Specialist (CKS) exam. Kubernetes is a Cloud Orchestration Platform providing reliability, replication, and stabilitywhile maximizing resource utilization for applications and services. By the conclusion of this hands-on, vendor agnostic training you will be equipped with a thorough understanding ofcloud security fundamentals, along with the knowledge, skills and abilities to secure a Kubernetes cluster, detect threats, and properly resolve a security catastrophe. This courseincludes hands-on instruction which develops skills and knowledge for securing container-based applications and Kubernetes platforms, during build, deployment, and runtime. We prioritizecovering all objectives and concepts necessary for passing the Certified Kubernetes Security Specialist (CKS) exam. You will be provided the components necessary to assemble your ownhigh availability Kubernetes environment and harden it for your security needs. Learning Your Environment Underlying Infrastructure Using Vim Tmux Cloud Security Primer Basic Principles Threat Analysis Approach CIS Benchmarks Securing your Kubernetes Cluster Kubernetes Architecture Pods and the Control Plane Kubernetes Security Concepts Install Kubernetes using kubeadm Configure Network Plugin Requirements Kubeadm Basic Cluster Installing Kubeadm Join Node to Cluster Kubeadm Token Manage Kubeadm Tokens Kubeadm Cluster Upgrade Securing the kube-apiserver Configuring the kube-apiserver Enable Audit Logging Falco Deploy Falco to Monitor System Calls Enable Pod Security Policies Encrypt Data at Rest Encryption Configuration Benchmark Cluster with Kube-Bench Kube-Bench Securing ETCD ETCD Isolation ETCD Disaster Recovery ETCD Snapshot and Restore Purge Kubernetes Purge Kubeadm 3Purge Kubeadm Image Scanning Container Essentials Secure Containers Creating a Docker Image Scanning with Trivy Trivy Snyk Security Manually Installing Kubernetes Kubernetes the Alta3 Way Deploy Kubernetes the Alta3 Way Validate your Kubernetes Installation Sonobuoy K8s Validation Test Kubectl (Optional) Kubectl get and sorting kubectl get kubectl describe Labels (Optional) Labels Labels and Selectors Annotations Insert an Annotation Securing your Application Scan a Running Container Tracee Security Contexts for Pods Understanding Security Contexts AppArmor Profiles AppArmor Isolate Container Kernels gVisor Pod Security Pod Security Policies Deploy a PSP Pod Security Standards Enable PSS Open Policy Agent (OPA) Admission Controller Create a LimitRange Open Policy Agent Policy as Code Deploy Gatekeeper User Administration Contexts Contexts Authentication and Authorization Role Based Access Control Role Based Access Control RBAC Distributing Access Service Accounts Limit Pod Service Accounts Securing Secrets Secrets Create and Consume Secrets Hashicorp Vault Deploy Vault Securing the Network Networking Plugins NetworkPolicy Deploy a NetworkPolicy mTLS Linkerd mTLS with istio istio Threat Detection Active Threat Analysis Host Intrusion Detection Deploy OSSEC Network Intrusion Detection Deploy Suricata Physical Intrusion Detection Disaster Recovery Harsh Reality of Security Deploy a Response Plan Kasten K10 Backups Deploy K10
Duration 5 Days 30 CPD hours This course is intended for This course is designed for Java developers who want to learn more about the specifications that comprise the world of Java Enterprise Edition (Java EE). Overview As a result of attending this course, you should be able to describe most of the specifications in Java EE 7 and create a component with each specification. You will be able to convert a Java SE program into a multi-tiered Java EE application. You should be able to demonstrate these skills: Describe the architecture of multi-tiered Java EE applications. Package Java EE applications and deploy to Red Hat JBoss Enterprise Application Platform with various tools. Create an Enterprise Java Bean instance. Manage the persistence of data using Java Persistence API. Create a web service using JAX-RS. Properly apply context scopes to beans and inject resources into Java Beans. Store and retrieve messages using the Java Messaging Service. Secure a Java EE application. Red Hat Application Development I: Programming in Java EE with exam (AD184) exposes experienced Java Standard Edition (Java SE) developers to the world of Java Enterprise Edition (Java EE). This course is based on Red Hat© Enterprise Application Platform 7.0. This course is a combination of Red Hat Application Development I: Programming in Java EE (AD183) and Red Hat Certified Enterprise Application Developer Exam (EX183). In this course, you will learn about the various specifications that make up Java EE. Through hands-on labs, you will transform a simple Java SE command line application into a multi-tiered enterprise application using various Java EE specifications, including Enterprise Java Beans, Java Persistence API, Java Messaging Service, JAX-RS for REST services, Contexts and Dependency Injection (CDI), and JAAS for securing the application. Transition to multi-tiered applications Describe Java EE features and distinguish between Java EE and Java SE applications. Package and deploying applications to an application server Describe the architecture of a Java EE application server, package an application, and deploy the application to an EAP server. Create Enterprise Java Beans Develop Enterprise Java Beans, including message-driven beans. Manage persistence Create persistence entities with validations. Manage entity relationships Define and manage JPA entity relationships. Create REST services Create REST APIs using the JAX-RS specification. Implement Contexts and Dependency Injection Describe typical use cases for using CDI and successfully implement it in an application. Create messaging applications with JMS Create messaging clients that send and receive messages using the JMS API. Secure Java EE applications Use JAAS to secure a Java EE application. Comprehensive review of Red Hat JBoss Development I: Java EE Demonstrate proficiency of the knowledge and skills obtained during the course. Additional course details: Nexus Humans Red Hat Application Development I: Programming in Java EE with exam (AD184) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Red Hat Application Development I: Programming in Java EE with exam (AD184) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 3 Days 18 CPD hours Overview In this course you?ll learn how to: Containerize and deploy a new Python script Configure the deployment with ConfigMaps, Secrets and SecurityContexts Understand multi-container pod design Configure probes for pod health Update and roll back an application Implement services and NetworkPolicies Use PersistentVolumeClaims for state persistence And more In this vendor agnostic course, you will use Python to build, monitor and troubleshoot scalable applications in Kubernetes. Introduction Objectives Who You Are The Linux Foundation Linux Foundation Training Preparing Your System Course Registration Labs Kubernetes Architecture What Is Kubernetes? Components of Kubernetes Challenges The Borg Heritage Kubernetes Architecture Terminology Master Node Minion (Worker) Nodes Pods Services Controllers Single IP per Pod Networking Setup CNI Network Configuration File Pod-to-Pod Communication Cloud Native Computing Foundation Resource Recommendations Labs Build Container Options Containerizing an Application Hosting a Local Repository Creating a Deployment Running Commands in a Container Multi-Container Pod readinessProbe livenessProbe Testing Labs Design Traditional Applications: Considerations Decoupled Resources Transience Flexible Framework Managing Resource Usage Multi-Container Pods Sidecar Container Adapter Container Ambassador Points to Ponder Labs Deployment Configuration Volumes Overview Introducing Volumes Volume Spec Volume Types Shared Volume Example Persistent Volumes and Claims Persistent Volume Persistent Volume Claim Dynamic Provisioning Secrets Using Secrets via Environment Variables Mounting Secrets as Volumes Portable Data with ConfigMaps Using ConfigMaps Deployment Configuration Status Scaling and Rolling Updates Deployment Rollbacks Jobs Labs Security Security Overview Accessing the API Authentication Authorization ABAC RBAC RBAC Process Overview Admission Controller Security Contexts Pod Security Policies Network Security Policies Network Security Policy Example Default Policy Example Labs Exposing Applications Service Types Services Diagram Service Update Pattern Accessing an Application with a Service Service without a Selector ClusterIP NodePort LoadBalancer ExternalName Ingress Resource Ingress Controller Labs Troubleshooting Troubleshotting Overview Basic Troubleshooting Steps Ongoing (Constant) Change Basic Troubleshooting Flow: Pods Basic Troubleshooting Flow: Node and Security Basic Troubleshooting Flow: Agents Monitoring Logging Tools Monitoring Applications System and Agent Logs Conformance Testing More Resource Labs Additional course details: Nexus Humans Kubernetes for App Developers training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Kubernetes for App Developers course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.