Duration 2 Days 12 CPD hours This course is intended for This introductory-level course is intended for Business Analysts and Data Analysts (or anyone else in the data science realm) who are already comfortable working with numerical data in Excel or other spreadsheet environments. No prior programming experience is required, and a browser is the only tool necessary for the course. Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Throughout the hands-on course students, will learn to leverage Python scripting for data science (to a basic level) using the most current and efficient skills and techniques. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore (to a basic level): How to work with Python interactively in web notebooks The essentials of Python scripting Key concepts necessary to enter the world of Data Science via Python This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it?s often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it's often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. Additional course details: Nexus Humans Python for Data Science Primer: Hands-on Technical Overview (TTPS4872) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Python for Data Science Primer: Hands-on Technical Overview (TTPS4872) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 5 Days 30 CPD hours This course is intended for This course is designed for application developers. Overview Deploy microservice applications on Red Hat© OpenShift Container Platform. Build a microservice application with Quarkus. Implement unit and integration tests for microservices. Use the config specification to inject data into a microservice. Secure a microservice using OAuth. Build and deploy native Quarkus applications. Develop microservice-based applications with Quarkus and OpenShift. Many enterprises are looking for a way to take advantage of cloud-native architectures, but many do not know the best approach. Quarkus is an exciting new technology that brings the reliability, familiarity, and maturity of Java Enterprise with a container-ready lightning fast deployment time. Red Hat Cloud-native Microservices Development with Quarkus (DO378) emphasizes learning architectural principles and implementing microservices based on Quarkus and OpenShift. You will build on application development fundamentals and focus on how to develop, monitor, test, and deploy modern microservices applications.. This course is based on OpenShift 4.5 and Quarkus 1.7L 1 - Describe microservice architectures Describe components and patterns of microservice-based application architectures. 2 - Implement a microservice with Quarkus Describe the specifications in Quarkus, implement a microservice with some of the specifications, and deploy it to an OpenShift cluster. 3 - Build microservice applications with Quarkus Build a persistent and configurable distributed quarkus microservices application. 4 - Implement fault tolerance Implement fault tolerance in a microservice architecture. 5 - Build and deploy native Quarkus applications Describe Quarkus in native mode and describe its deployment on OpenShift Container Platform. 6 - Test microservices Implement unit and integration tests for microservices. 7 - Create application health checks Create a health check for a microservice. 8 - Secure microservices Secure microservice endpoints and communication. 9 - Monitor microservices Monitor the operation of a microservice using metrics and distributed tracing. Additional course details: Nexus Humans Red Hat Cloud-native Microservices Development with Quarkus (DO378) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Red Hat Cloud-native Microservices Development with Quarkus (DO378) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 4 Days 24 CPD hours This course is intended for This class is intended for experienced developers who are responsible for managing big data transformations including: Extracting, loading, transforming, cleaning, and validating data. Designing pipelines and architectures for data processing. Creating and maintaining machine learning and statistical models. Querying datasets, visualizing query results and creating reports Overview Design and build data processing systems on Google Cloud Platform. Leverage unstructured data using Spark and ML APIs on Cloud Dataproc. Process batch and streaming data by implementing autoscaling data pipelines on Cloud Dataflow. Derive business insights from extremely large datasets using Google BigQuery. Train, evaluate and predict using machine learning models using TensorFlow and Cloud ML. Enable instant insights from streaming data Get hands-on experience with designing and building data processing systems on Google Cloud. This course uses lectures, demos, and hand-on labs to show you how to design data processing systems, build end-to-end data pipelines, analyze data, and implement machine learning. This course covers structured, unstructured, and streaming data. Introduction to Data Engineering Explore the role of a data engineer. Analyze data engineering challenges. Intro to BigQuery. Data Lakes and Data Warehouses. Demo: Federated Queries with BigQuery. Transactional Databases vs Data Warehouses. Website Demo: Finding PII in your dataset with DLP API. Partner effectively with other data teams. Manage data access and governance. Build production-ready pipelines. Review GCP customer case study. Lab: Analyzing Data with BigQuery. Building a Data Lake Introduction to Data Lakes. Data Storage and ETL options on GCP. Building a Data Lake using Cloud Storage. Optional Demo: Optimizing cost with Google Cloud Storage classes and Cloud Functions. Securing Cloud Storage. Storing All Sorts of Data Types. Video Demo: Running federated queries on Parquet and ORC files in BigQuery. Cloud SQL as a relational Data Lake. Lab: Loading Taxi Data into Cloud SQL. Building a Data Warehouse The modern data warehouse. Intro to BigQuery. Demo: Query TB+ of data in seconds. Getting Started. Loading Data. Video Demo: Querying Cloud SQL from BigQuery. Lab: Loading Data into BigQuery. Exploring Schemas. Demo: Exploring BigQuery Public Datasets with SQL using INFORMATION_SCHEMA. Schema Design. Nested and Repeated Fields. Demo: Nested and repeated fields in BigQuery. Lab: Working with JSON and Array data in BigQuery. Optimizing with Partitioning and Clustering. Demo: Partitioned and Clustered Tables in BigQuery. Preview: Transforming Batch and Streaming Data. Introduction to Building Batch Data Pipelines EL, ELT, ETL. Quality considerations. How to carry out operations in BigQuery. Demo: ELT to improve data quality in BigQuery. Shortcomings. ETL to solve data quality issues. Executing Spark on Cloud Dataproc The Hadoop ecosystem. Running Hadoop on Cloud Dataproc. GCS instead of HDFS. Optimizing Dataproc. Lab: Running Apache Spark jobs on Cloud Dataproc. Serverless Data Processing with Cloud Dataflow Cloud Dataflow. Why customers value Dataflow. Dataflow Pipelines. Lab: A Simple Dataflow Pipeline (Python/Java). Lab: MapReduce in Dataflow (Python/Java). Lab: Side Inputs (Python/Java). Dataflow Templates. Dataflow SQL. Manage Data Pipelines with Cloud Data Fusion and Cloud Composer Building Batch Data Pipelines visually with Cloud Data Fusion. Components. UI Overview. Building a Pipeline. Exploring Data using Wrangler. Lab: Building and executing a pipeline graph in Cloud Data Fusion. Orchestrating work between GCP services with Cloud Composer. Apache Airflow Environment. DAGs and Operators. Workflow Scheduling. Optional Long Demo: Event-triggered Loading of data with Cloud Composer, Cloud Functions, Cloud Storage, and BigQuery. Monitoring and Logging. Lab: An Introduction to Cloud Composer. Introduction to Processing Streaming Data Processing Streaming Data. Serverless Messaging with Cloud Pub/Sub Cloud Pub/Sub. Lab: Publish Streaming Data into Pub/Sub. Cloud Dataflow Streaming Features Cloud Dataflow Streaming Features. Lab: Streaming Data Pipelines. High-Throughput BigQuery and Bigtable Streaming Features BigQuery Streaming Features. Lab: Streaming Analytics and Dashboards. Cloud Bigtable. Lab: Streaming Data Pipelines into Bigtable. Advanced BigQuery Functionality and Performance Analytic Window Functions. Using With Clauses. GIS Functions. Demo: Mapping Fastest Growing Zip Codes with BigQuery GeoViz. Performance Considerations. Lab: Optimizing your BigQuery Queries for Performance. Optional Lab: Creating Date-Partitioned Tables in BigQuery. Introduction to Analytics and AI What is AI?. From Ad-hoc Data Analysis to Data Driven Decisions. Options for ML models on GCP. Prebuilt ML model APIs for Unstructured Data Unstructured Data is Hard. ML APIs for Enriching Data. Lab: Using the Natural Language API to Classify Unstructured Text. Big Data Analytics with Cloud AI Platform Notebooks What's a Notebook. BigQuery Magic and Ties to Pandas. Lab: BigQuery in Jupyter Labs on AI Platform. Production ML Pipelines with Kubeflow Ways to do ML on GCP. Kubeflow. AI Hub. Lab: Running AI models on Kubeflow. Custom Model building with SQL in BigQuery ML BigQuery ML for Quick Model Building. Demo: Train a model with BigQuery ML to predict NYC taxi fares. Supported Models. Lab Option 1: Predict Bike Trip Duration with a Regression Model in BQML. Lab Option 2: Movie Recommendations in BigQuery ML. Custom Model building with Cloud AutoML Why Auto ML? Auto ML Vision. Auto ML NLP. Auto ML Tables.
Duration 3.75 Days 22.5 CPD hours This course is intended for New systems administrators Overview When you complete this course, you will be able to:Customize your application, including page layouts, fields, tabs, and business processes in Lightning Experience.Learn how security settings created in Salesforce Classic are applied in Lightning.Maintain and import clean data in Lightning.Use Lightning features to create high-value reports and dashboards.Understand how workflow automation complies with Lightning This course is the core training that ensures your success with Salesforce Lightning. It?s a must for new administrators, and we recommend completing this course before starting a Salesforce deployment or when taking over an existing deployment. Getting Around the App Data Model and Navigation Lightning Experience Help & Training Getting Your Organization Ready for Users Setting Up the Company Profile Configuring the User Interface Setting Up Activities and Calendars Configuring Search Settings Setting Up Chatter Groups Mobile Access with Salesforce1 Setting Up & Managing Users Managing User Profiles Managing Users Setting Up Chatter Free Users and Invites Troubleshooting Login Issues Security & Data Access Restricting Logins Determining Object Access Setting Up Record Access Creating a Role Hierarchy Dealing with Record Access Exceptions Managing Field-level Security Object Customizations Administering Standard Fields Creating New Custom Fields Creating Selection Fields: Picklists and Lookups Creating Formula Fields Working with Page Layouts Working with Record Types and Business Processes Maintaining Data Quality Managing Data Import Wizards Data Loader Data.com Mass Transfer Backing Up Data Mass Delete and Recycle Bin Reports & Dashboards Running and Modifying Reports Creating New Reports with the Report Builder Working with Report Filters Summarizing with Formulas and Visual Summaries Printing, Exporting, and Emailing Reports Building Dashboards Automation Email Templates Workflow Rules Process Builder Lead Automation Managing the Support Process Managing and Resolving Cases Customizing a Support Process Automating Support Understanding the Salesforce Console for Service Collaborating in the Service Cloud Analyzing Support Data Additional course details: Nexus Humans Salesforce Administration Essentials for New Admins in Lightning Experience (ADX201) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Salesforce Administration Essentials for New Admins in Lightning Experience (ADX201) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
The main topics to be covered in this course are about everyday activities, family, friends, life events, remembering and forgetting, characteristics of people, holidays and more; Students practise their speaking, writing, listening and reading skills in and out of class with their teachers and classmates using their books and applications; The course is based on Ελληνικά Α and Κλικ στα Ελληνικά. In addition, there is a wide range of language material designed by the Greek LOL Teachers. All lessons are supported by educational applications to help students enrich their vocabulary and practise their Greek with fun and interactive exercises, videos and games.
Duration 1 Days 6 CPD hours This course is intended for This course is designed for a non-technical audience and doesn't require any prior coding or technical experience. The handson exercises will be done using pre-built OpenAI tools and interfaces that are user-friendly and easy to use. Overview Working in an interactive learning environment, led by our engaging expert, you will: Get comfortable with the basics of prompt engineering and discover how it can make a difference in various business tasks, such as enhancing customer support, creating content, and fine-tuning sales pitches. Develop the knack for crafting, refining, and perfecting prompts suited to specific business situations by understanding context, user intent, and what makes a prompt great. Learn how to smoothly incorporate prompt engineering solutions into your existing business workflows, including pinpointing the right processes, integrating with your current software, and keeping data privacy and security in check. Become proficient in advanced techniques and best practices in prompt engineering, like making use of APIs, customizing language models, and collaborating with your teammates across different departments. Keep up with the latest developments in prompt engineering and be ready to adapt to changing business needs and trends, ensuring that you stay relevant and continue to grow in the dynamic business world. Prompt engineering is the process of designing and refining input prompts to get desired output from advanced language models, such as OpenAI Codex or GPT-4. It involves creating effective questions or statements that guide the AI model to generate useful responses for a specific task or purpose, like enhancing customer support, generating content, and fine-tuning sales pitches, making it an essential skill set for a wide range of business applications. Quick Start to Prompt Engineering for Everyday Business Users is a one-day, workshop style hands-on course that where you'll learn how to create effective prompts, integrate prompt engineering solutions into existing workflows, and uncover advanced techniques and best practices. Guided by our engaging, expert instructor, you?ll experiment with innovative tools and develop practical skills that can be immediately applied to a variety of projects. Whether you're aiming to enhance customer interactions, simplify content creation, or refine internal communication, this immersive learning experience will equip you with the knowledge to make a meaningful impact on your organization. Introduction to Prompt Engineering Understand the fundamentals of prompt engineering and its applications in the business world. What is prompt engineering? Importance of prompt engineering in business Key concepts and terminology Examples of prompt engineering in business scenarios Overview of popular prompt engineering tools (e.g., OpenAI Codex, GPT-4) Activity: Hands-on exploration of prompt engineering tools: Participants will engage in a fun scavenger hunt activity, where they will experiment with different prompt engineering tools to answer a set of questions. Developing Effective Prompts Learn how to create and refine prompts for a variety of business applications. Anatomy of a good prompt Understanding context and user intent Techniques for prompt iteration and optimization Generating specific and creative responses Handling sensitive information and biases Activity: Prompt development workshop: Participants will practice developing and refining prompts in a collaborative, game-like environment, where they will compete to create the most effective prompts for given business scenarios. Integrating Prompt Engineering into Business Processes Discover how to incorporate prompt engineering solutions into existing workflows. Identifying business processes that can benefit from prompt engineering Integrating prompt engineering with existing software and tools Evaluating the success and impact of prompt engineering solutions Ensuring data privacy and security Scaling prompt engineering solutions across an organization Activity: Business process integration simulation: Participants will work in teams to create a plan for integrating a prompt engineering solution into a simulated business process, with a focus on creativity and practicality. Advanced Techniques and Best Practices Gain insights into advanced techniques and best practices for prompt engineering in a business context. Leveraging APIs for prompt engineering Customizing and fine-tuning language models Adapting to changing business requirements and trends Collaborating with cross-functional teams Staying up-to-date with prompt engineering advancements Activity: Advanced prompt engineering challenge: Participants will take part in a friendly competition, using advanced techniques to solve complex business-related prompt engineering challenges. Additional course details: Nexus Humans QuickStart to Prompt Engineering for Everyday Business Users (TTAI2009) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the QuickStart to Prompt Engineering for Everyday Business Users (TTAI2009) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Looking to master Linux? Explore IPSR's comprehensive Linux online training for a robust learning experience. Our meticulously designed Linux online training course covers everything, ensuring a deep understanding of Linux systems, commands, and administration. With IPSR's Linux online training, dive into the world of open-source mastery, learning from industry experts at your convenience. Gain hands-on experience and practical skills in Linux administration, scripting, security, and more. Our tailored curriculum and interactive sessions empower learners to tackle real-world challenges confidently. Join IPSR's Linux online training to unleash your potential and become proficient in Linux, setting yourself on a path to career success in the tech industry.
Duration 2 Days 12 CPD hours This course is intended for Data Modelers Participants will learn the full scope of the metadata modeling process, from initial project creation, to publishing a dynamic cube, and enabling end users to easily author reports and analyze data. Introduction to IBM Cognos Dynamic Cubes Define and differentiate Dynamic Cubes Dynamic Cubes characteristics Examine Dynamic Cube requirements Examine Dynamic Cube components Examine high level architecture IBM Cognos Dynamic Query Review Dimensional Data Structures Dynamic Cubes caching Create & Design a Dynamic Cube Explore the IBM Cognos Cube Designer Review the cube development process Examine the Automatic Cube Generation Manual development overview Create dimensions Model the cube Best practice for effective modeling Deploy & Configure a Dynamic Cube Deploy a cube Explore the Estimate Hardware Requirements Identify cube management tasks Examine Query Service administration Explore Dynamic Cube properties Schedule cube actions Use the DCAdmin comment line tool Advanced Dynamic Cube Modelling Examine advanced modeling concepts Explore modeling caveats Calculated measures and members Model Relative Time Explore the Current Period property Define period aggregation rules for measures Advanced Features of Cube Designer Examine multilingual support Examine ragged hierarchies and padding members Define Parent-Child Dimensions Refresh Metadata Import Framework Manager packages Filter measures and dimensions Optimize Performance with Aggregates Identify aggregates and aggregate tables In-memory aggregates Use Aggregate Advisor to identify aggregates User defined in-memory aggregates Optimize In-Memory Aggregates automatically Aggregate Advisor recommendations Monitor Dynamic Cube performance Model aggregates (automatically vs manually) Use Slicers to define aggregation partitions Define Security Overview of Dynamic Cube security Identify security filters The Security process - Three steps Examine security scope Identify scope rules Identify roles Capabilities and access permissions Cube security deep dive Model a Virtual Cube Explore virtual cubes Create the virtual cube Explore virtual cube objects Examine virtual measures and calculated members Currency conversion using virtual cubes Security on virtual cubes Introduction to IBM Cognos Analytics Define IBM Cognos Analytics Redefined Business Intelligence Self-service Navigate to content in IBM Cognos Analytics Interact with the user interface Model data with IBM Cognos Analytics IBM Cognos Analytics components Create reports Perform self-service with analysis and Dashboards IBM Cognos Analytics architecture (high level) IBM Cognos Analytics security Package / data source relationship Create Data modules Upload files Additional course details: Nexus Humans B6063 IBM Cognos Cube Designer - Design Dynamic Cubes (v11.0) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the B6063 IBM Cognos Cube Designer - Design Dynamic Cubes (v11.0) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Qualification Number 600/3775/1 Minimum entry age 19 Guided Learning Hours 55 What does this qualification cover? This qualification is designed to provide learners with the knowledge and understanding required to successfully lead a team. Who is it suitable for? This qualification is for those individuals who wish to develop their knowledge and understanding of leading a team and / or are planning to move into a team leading role This qualification is suitable for learners aged 19 and above. What are the entry requirements? There are no specific recommended prior learning requirements for this qualification. How is this qualification structured? The qualification is made up of three mandatory units: • Unit 1 Team Leading Responsibilities • Unit 2 Managing yourself and providing direction to your team • Unit 3 Communicating Information and Knowledge in the Workplace How is it assessed? This qualification is assessed via an internally assessed and externally verified portfolio of evidence. Funding for our qualifications This qualification is available fully funded under the Adult Education Budget funding stream and is free to the candidate.
Duration 3 Days 18 CPD hours Overview In this course you?ll learn how to: Containerize and deploy a new Python script Configure the deployment with ConfigMaps, Secrets and SecurityContexts Understand multi-container pod design Configure probes for pod health Update and roll back an application Implement services and NetworkPolicies Use PersistentVolumeClaims for state persistence And more In this vendor agnostic course, you will use Python to build, monitor and troubleshoot scalable applications in Kubernetes. Introduction Objectives Who You Are The Linux Foundation Linux Foundation Training Preparing Your System Course Registration Labs Kubernetes Architecture What Is Kubernetes? Components of Kubernetes Challenges The Borg Heritage Kubernetes Architecture Terminology Master Node Minion (Worker) Nodes Pods Services Controllers Single IP per Pod Networking Setup CNI Network Configuration File Pod-to-Pod Communication Cloud Native Computing Foundation Resource Recommendations Labs Build Container Options Containerizing an Application Hosting a Local Repository Creating a Deployment Running Commands in a Container Multi-Container Pod readinessProbe livenessProbe Testing Labs Design Traditional Applications: Considerations Decoupled Resources Transience Flexible Framework Managing Resource Usage Multi-Container Pods Sidecar Container Adapter Container Ambassador Points to Ponder Labs Deployment Configuration Volumes Overview Introducing Volumes Volume Spec Volume Types Shared Volume Example Persistent Volumes and Claims Persistent Volume Persistent Volume Claim Dynamic Provisioning Secrets Using Secrets via Environment Variables Mounting Secrets as Volumes Portable Data with ConfigMaps Using ConfigMaps Deployment Configuration Status Scaling and Rolling Updates Deployment Rollbacks Jobs Labs Security Security Overview Accessing the API Authentication Authorization ABAC RBAC RBAC Process Overview Admission Controller Security Contexts Pod Security Policies Network Security Policies Network Security Policy Example Default Policy Example Labs Exposing Applications Service Types Services Diagram Service Update Pattern Accessing an Application with a Service Service without a Selector ClusterIP NodePort LoadBalancer ExternalName Ingress Resource Ingress Controller Labs Troubleshooting Troubleshotting Overview Basic Troubleshooting Steps Ongoing (Constant) Change Basic Troubleshooting Flow: Pods Basic Troubleshooting Flow: Node and Security Basic Troubleshooting Flow: Agents Monitoring Logging Tools Monitoring Applications System and Agent Logs Conformance Testing More Resource Labs Additional course details: Nexus Humans Kubernetes for App Developers training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Kubernetes for App Developers course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.