Duration 2 Days 12 CPD hours This course is intended for This intermediate course is for Business and Technical Specialist working with the Matching, Linking, and Search services of InfoSphere MDM Virtual module. Overview Understand how Matching and Linking work for both the Virtual Implementations of InfoSphere MDM Understand the MDM configuration project and database tables used by the PME Understand the PME Algorithms (Standardization, Bucketing and Comparison steps) and how to create and customize the algorithms using the workbench Understand how to analyze the Bucketing steps in an algorithm Understand how to generate weights for a given algorithm and how those weights are generated based on a sample database set Understand how to analyze the weights that are generated using the workbench Understand how to deploy the PME configuration for the Virtual implementations of InfoSphere MDM The InfoSphere MDM Virtual Module Algorithms V.11 course prepares students to work with and customize the algorithm configurations deployed to the InfoSphere MDM Probabilistic Matching Engine (PME) for Virtual MDM implementations. PME and Virtual Overview Virtual MDM Overview Terminology (Source, Entity, Member, Attributes) PME and Virtual MDM ( Algorithms, Weights, Comparison Scores, Thresholds) Virtual MDM Linkages and Tasks Virtual MDM Algorithms Standardization Bucketing Comparison Functions Virtual PME Data Model Algorithm configuration tables Member Derived Data Bucketing Data Bucket Analysis Analysis Overview Attribute Completeness Bucket Analysis Weights Weights Overview (Frequency-based weights, Edit Distance weights and Parameterize weights) The weight formula Running weight generation Analyzing weights Bulk Cross Match process Pair Manager Threshold calculations Additional course details: Nexus Humans ZZ880 IBM Virtual Module Algorithms for InfoSphere MDM V11 training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the ZZ880 IBM Virtual Module Algorithms for InfoSphere MDM V11 course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: A technical audience at an intermediate level Overview Using Amazon SageMaker, this course teaches you how to: Prepare a dataset for training. Train and evaluate a machine learning model. Automatically tune a machine learning model. Prepare a machine learning model for production. Think critically about machine learning model results In this course, learn how to solve a real-world use case with machine learning and produce actionable results using Amazon SageMaker. This course teaches you how to use Amazon SageMaker to cover the different stages of the typical data science process, from analyzing and visualizing a data set, to preparing the data and feature engineering, down to the practical aspects of model building, training, tuning and deployment. Day 1 Business problem: Churn prediction Load and display the dataset Assess features and determine which Amazon SageMaker algorithm to use Use Amazon Sagemaker to train, evaluate, and automatically tune the model Deploy the model Assess relative cost of errors Additional course details: Nexus Humans Practical Data Science with Amazon SageMaker training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Practical Data Science with Amazon SageMaker course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 5 Days 30 CPD hours This course is intended for This introductory-level, fast-paced course is for skilled web developers new to React who have prior experienced working HTML5, CSS3 and JavaScript. Overview Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore: A basic and advanced understanding of React components An advanced, in-depth knowledge of how React works A complete understanding of using Redux How to build, validate, and populate interactive forms How to use inline styles for perfect looking components How to test React components How to build and use components How to get control of your build process A deep understanding of data-driven modeling with props and state How to use client-side routing for pages in your apps How to debug a React application Mastering React is a comprehensive hands-on course that aims to be the single most useful resource on getting up to speed quickly with React. Geared for more experienced web developers new to React, this course provides students with the core knowledge and hands-on skills they require to build reliable, powerful React apps. After the first few modules, you?ll have a solid understanding of React?s fundamentals and will be able to build a wide array of rich, interactive web apps with the framework. The first module is an introduction to the new functionality in ECMAScript 6 (JavaScript). Client-side routing between pages, managing complex state, and heavy API interaction at scale are also covered. This course consists of two parts. In the first part of the course students will explore all the fundamentals with a progressive, example-driven approach. You?ll create your first apps, learn how to write components, start handling user interaction, and manage rich forms. We end the first part by exploring the inner workings of Create React App (Facebook?s tool for running React apps), writing automated unit tests, and building a multi-page app that uses client-side routing. The latter part of the course moves into more advanced concepts that you?ll see used in large, production applications. These concepts explore strategies for data architecture, transport, and management: Redux is a state management paradigm based on the Flux architecture. Redux provides a structure for large state trees and allows you to decouple user interaction in your app from state changes. GraphQL is a powerful, typed, REST API alternative where the client describes the data it needs. Hooks is the powerful, new way to maintain state and properties with functional components and the future of React according to Facebook. ES6 Primer (Optional) Prefer const and let over var Arrow functions Modules Object.assign() Template literals The spread operator and Rest parameters Enhanced object literals Default arguments Destructuring assignments Your first React Web Application Setting up your development environment JavaScript ES6 /ES7 Getting started What?s a component? Our first component Building the App Making the App data-driven Your app?s first interaction Updating state and immutability Refactoring with the Babel plugin transform-class-properties JSX and the Virtual DOM React Uses a Virtual DOM Why Not Modify the Actual DOM? What is a Virtual DOM? Virtual DOM Pieces ReactElement JSX JSX Creates Elements JSX Attribute Expressions JSX Conditional Child Expressions JSX Boolean Attributes JSX Comments JSX Spread Syntax JSX Gotchas JSX Summary Components A time-logging app Getting started Breaking the app into components The steps for building React apps from scratch Updating timers Deleting timers Adding timing functionality Add start and stop functionality Methodology review Advanced Component Configuration with props, state, and children ReactComponent props are the parameters PropTypes Default props with getDefaultProps() context state Stateless Components Talking to Children Components with props.children Forms Forms 101 Text Input Remote Data Async Persistence Redux Form Modules Unit Testing & Jest Writing tests without a framework What is Jest? Using Jest Testing strategies for React applications Testing a basic React component with Enzyme Writing tests for the food lookup app Writing FoodSearch.test.js Routing What?s in a URL? React Router?s core components Building the components of react-router Dynamic routing with React Router Supporting authenticated routes Intro to Flux and Redux Why Flux? Flux is a Design Pattern Flux implementations Redux & Redux?s key ideas Building a counter The core of Redux The beginnings of a chat app Building the reducer() Subscribing to the store Connecting Redux to React Intermediate Redux Using createStore() from the redux library Representing messages as objects in state Introducing threads Adding the ThreadTabs component Supporting threads in the reducer Adding the action OPEN_THREAD Breaking up the reducer function Adding messagesReducer() Defining the initial state in the reducers Using combineReducers() from redux React Hooks Motivation behind Hooks How Hooks Map to Component Classes Using Hooks Requires react 'next' useState() Hook Example useEffect() Hook Example useContext() Hook Example Using Custom Hooks Using Webpack with Create React App JavaScript modules Create React App Exploring Create React App Webpack basics Making modifications Hot reloading; Auto-reloading Creating a production build Ejecting Using Create React App with an API server When to use Webpack/Create React App Using GraphQL Your First GraphQL Query GraphQL Benefits GraphQL vs. REST GraphQL vs. SQL Relay and GraphQL Frameworks Chapter Preview Consuming GraphQL Exploring With GraphiQL GraphQL Syntax 101 . Complex Types Exploring a Graph Graph Nodes ; Viewer Graph Connections and Edges Mutations Subscriptions GraphQL With JavaScript GraphQL With React
Duration 2 Days 12 CPD hours This course is intended for This introductory-level course is intended for Business Analysts and Data Analysts (or anyone else in the data science realm) who are already comfortable working with numerical data in Excel or other spreadsheet environments. No prior programming experience is required, and a browser is the only tool necessary for the course. Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Throughout the hands-on course students, will learn to leverage Python scripting for data science (to a basic level) using the most current and efficient skills and techniques. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore (to a basic level): How to work with Python interactively in web notebooks The essentials of Python scripting Key concepts necessary to enter the world of Data Science via Python This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it?s often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. An Overview of Python Why Python? Python in the Shell Python in Web Notebooks (iPython, Jupyter, Zeppelin) Demo: Python, Notebooks, and Data Science Getting Started Using variables Builtin functions Strings Numbers Converting among types Writing to the screen Command line parameters Flow Control About flow control White space Conditional expressions Relational and Boolean operators While loops Alternate loop exits Sequences, Arrays, Dictionaries and Sets About sequences Lists and list methods Tuples Indexing and slicing Iterating through a sequence Sequence functions, keywords, and operators List comprehensions Generator Expressions Nested sequences Working with Dictionaries Working with Sets Working with files File overview Opening a text file Reading a text file Writing to a text file Reading and writing raw (binary) data Functions Defining functions Parameters Global and local scope Nested functions Returning values Essential Demos Sorting Exceptions Importing Modules Classes Regular Expressions The standard library Math functions The string module Dates and times Working with dates and times Translating timestamps Parsing dates from text Formatting dates Calendar data Python and Data Science Data Science Essentials Pandas Overview NumPy Overview SciKit Overview MatPlotLib Overview Working with Python in Data Science Additional course details: Nexus Humans Python for Data Science: Hands-on Technical Overview (TTPS4873) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Python for Data Science: Hands-on Technical Overview (TTPS4873) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Architecture and course review Post course knowledge check Architecture review Course review
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake architectures. Module 1: Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Module 2: Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Module 3: Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Module 4: Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Module 5: Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Module 6: Architecture and course review Post course knowledge check Architecture review Course review Additional course details: Nexus Humans Building Data Lakes on AWS training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Building Data Lakes on AWS course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours In this hands on workshop for Agile Scrum Masters, Release Train Engineers and anyone serving as Jira Administrators, Jira experts will lead you through advanced configuration and customization settings in Jira, from installation through to customized screens, workflows, filters and reports. Jira Administration Adding and managing Users Administering and managing Groups Global Jira Settings Jira layout and interface customization User authentication and security Jira Customization Customization of screens and fields Customization of workflows Project and Board Administration Configuring and managing Projects Configuring and managing Boards Creating and managing Filters JQL Jira Integration Integrating Jira with Atlassian Tools Retrospectives and Documentation in Confluence Code management with Bitbucket Integration management with Bamboo Building a Dashboard with gadgets Jira Plug-ins and Marketplace
Duration 4 Days 24 CPD hours This course is intended for This course is intended for: Developers Solutions Architects Data Engineers Anyone with little to no experience with ML and wants to learn about the ML pipeline using Amazon SageMaker Overview In this course, you will learn to: Select and justify the appropriate ML approach for a given business problem Use the ML pipeline to solve a specific business problem Train, evaluate, deploy, and tune an ML model using Amazon SageMaker Describe some of the best practices for designing scalable, cost-optimized, and secure ML pipelines in AWS Apply machine learning to a real-life business problem after the course is complete This course explores how to use the machine learning (ML) pipeline to solve a real business problem in a project-based learning environment. Students will learn about each phase of the pipeline from instructor presentations and demonstrations and then apply that knowledge to complete a project solving one of three business problems: fraud detection, recommendation engines, or flight delays. By the end of the course, students will have successfully built, trained, evaluated, tuned, and deployed an ML model using Amazon SageMaker that solves their selected business problem. Module 0: Introduction Pre-assessment Module 1: Introduction to Machine Learning and the ML Pipeline Overview of machine learning, including use cases, types of machine learning, and key concepts Overview of the ML pipeline Introduction to course projects and approach Module 2: Introduction to Amazon SageMaker Introduction to Amazon SageMaker Demo: Amazon SageMaker and Jupyter notebooks Hands-on: Amazon SageMaker and Jupyter notebooks Module 3: Problem Formulation Overview of problem formulation and deciding if ML is the right solution Converting a business problem into an ML problem Demo: Amazon SageMaker Ground Truth Hands-on: Amazon SageMaker Ground Truth Practice problem formulation Formulate problems for projects Module 4: Preprocessing Overview of data collection and integration, and techniques for data preprocessing and visualization Practice preprocessing Preprocess project data Class discussion about projects Module 5: Model Training Choosing the right algorithm Formatting and splitting your data for training Loss functions and gradient descent for improving your model Demo: Create a training job in Amazon SageMaker Module 6: Model Evaluation How to evaluate classification models How to evaluate regression models Practice model training and evaluation Train and evaluate project models Initial project presentations Module 7: Feature Engineering and Model Tuning Feature extraction, selection, creation, and transformation Hyperparameter tuning Demo: SageMaker hyperparameter optimization Practice feature engineering and model tuning Apply feature engineering and model tuning to projects Final project presentations Module 8: Deployment How to deploy, inference, and monitor your model on Amazon SageMaker Deploying ML at the edge Demo: Creating an Amazon SageMaker endpoint Post-assessment Course wrap-up Additional course details: Nexus Humans The Machine Learning Pipeline on AWS training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the The Machine Learning Pipeline on AWS course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 2 Days 12 CPD hours This course is intended for This course is intended for Business Leaders, including managers/supervisors in the following roles: Developer Architect Video Operator Overview In this course, you will learn to: Articulate the essential terms and concepts fundamental to video compression and distribution Describe the four fundamental stages of video streaming workflows: ingest, process, store and deliver Explain the importance of security in the AWS Cloud and how it is applied in video streaming workflows Analyze video streaming workflow diagrams using AWS services, based on simple to complex use cases Describe some of the key variables that influence workflow decisions Recognize how other AWS services for compliance, storage, and compute, interact with AWS Media Services in video streaming workflows and the functions they perform Describe strategies to test or prototype workflows to mitigate risk and cost impacts and optimize video streaming workflows Use the AWS Management Console to build and run simple video streaming workflows for live and video-on-demand content Recognize the automation and data analytics available for Media Services when used with AWS AI and explore media-specific use cases for these services Identify the next steps in exploring migration to the cloud for one or more Media Services This course covers the media and cloud fundamentals that will empower you to develop a cloud migration strategy for media workflows in support of business goals. The course covers important concepts related to video processing and delivery, the variables that can impact migration decisions, and real-world examples of hybrid and cloud use cases for AWS Media Services. It also introduces security, artificial intelligence, and analytics concepts to help you consider how AWS Media Services fit into your overall cloud strategy. Module 1: Important video concepts Video Metrics Video Compression Video Distribution Major Protocols Used in Video Streaming Module 2: Anatomy of streaming workflows Ingest Process Store Deliver Module 3: Using AWS services in media workflows video-on-demand (VOD) Introduction to AWS Media Services Security Variables Impacting Workflow Design VOD Simple Use Cases VOD Advanced Use Cases Lab 1: Build and run a simple video streaming workflow for VOD content Module 4: Using AWS services in media workflows live streaming Challenges of Live Streaming Live Streaming Simple Use Cases Live Streaming Advanced Use Cases Lab 2: Build and run a simple video streaming workflow for live content Module 5: Optimizing Workflows Cost Considerations Mitigating Risk Monitoring and Automation Exploring Migration Options Additional course details: Nexus Humans AWS Media Essentials for IT Business Decision Makers training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the AWS Media Essentials for IT Business Decision Makers course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 2 Days 12 CPD hours This course is intended for Cloud Solutions Architects, Site Reliability Engineers, Systems Operations professionals, DevOps Engineers, IT managers. Individuals using Google Cloud Platform to create new solutions or to integrate existing systems, application environments, and infrastructure with the Google Cloud Platform. Overview Apply a tool set of questions, techniques and design considerations Define application requirements and express them objectively as KPIs, SLO's and SLI's Decompose application requirements to find the right microservice boundaries Leverage Google Cloud developer tools to set up modern, automated deployment pipelines Choose the appropriate Google Cloud Storage services based on application requirements Architect cloud and hybrid networks Implement reliable, scalable, resilient applications balancing key performance metrics with cost Choose the right Google Cloud deployment services for your applications Secure cloud applications, data and infrastructure Monitor service level objectives and costs using Stackdriver tools This course features a combination of lectures, design activities, and hands-on labs to show you how to use proven design patterns on Google Cloud to build highly reliable and efficient solutions and operate deployments that are highly available and cost-effective. This course was created for those who have already completed the Architecting with Google Compute Engine or Architecting with Google Kubernetes Engine course. Defining the Service Describe users in terms of roles and personas. Write qualitative requirements with user stories. Write quantitative requirements using key performance indicators (KPIs). Evaluate KPIs using SLOs and SLIs. Determine the quality of application requirements using SMART criteria. Microservice Design and Architecture Decompose monolithic applications into microservices. Recognize appropriate microservice boundaries. Architect stateful and stateless services to optimize scalability and reliability. Implement services using 12-factor best practices. Build loosely coupled services by implementing a well-designed REST architecture. Design consistent, standard RESTful service APIs. DevOps Automation Automate service deployment using CI/CD pipelines. Leverage Cloud Source Repositories for source and version control. Automate builds with Cloud Build and build triggers. Manage container images with Google Container Registry. Create infrastructure with code using Deployment Manager and Terraform. Choosing Storage Solutions Choose the appropriate Google Cloud data storage service based on use case, durability, availability, scalability and cost. Store binary data with Cloud Storage. Store relational data using Cloud SQL and Spanner. Store NoSQL data using Firestore and Cloud Bigtable. Cache data for fast access using Memorystore. Build a data warehouse using BigQuery. Google Cloud and Hybrid Network Architecture Design VPC networks to optimize for cost, security, and performance. Configure global and regional load balancers to provide access to services. Leverage Cloud CDN to provide lower latency and decrease network egress. Evaluate network architecture using the Cloud Network Intelligence Center. Connect networks using peering and VPNs. Create hybrid networks between Google Cloud and on-premises data centers using Cloud Interconnect. Deploying Applications to Google Cloud Choose the appropriate Google Cloud deployment service for your applications. Configure scalable, resilient infrastructure using Instance Templates and Groups. Orchestrate microservice deployments using Kubernetes and GKE. Leverage App Engine for a completely automated platform as a service (PaaS). Create serverless applications using Cloud Functions. Designing Reliable Systems Design services to meet requirements for availability, durability, and scalability. Implement fault-tolerant systems by avoiding single points of failure, correlated failures, and cascading failures. Avoid overload failures with the circuit breaker and truncated exponential backoff design patterns. Design resilient data storage with lazy deletion. Analyze disaster scenarios and plan for disaster recovery using cost/risk analysis. Security Design secure systems using best practices like separation of concerns, principle of least privilege, and regular audits. Leverage Cloud Security Command Center to help identify vulnerabilities. Simplify cloud governance using organizational policies and folders. Secure people using IAM roles, Identity-Aware Proxy, and Identity Platform. Manage the access and authorization of resources by machines and processes using service accounts. Secure networks with private IPs, firewalls, and Private Google Access. Mitigate DDoS attacks by leveraging Cloud DNS and Cloud Armor. Maintenance and Monitoring Manage new service versions using rolling updates, blue/green deployments, and canary releases. Forecast, monitor, and optimize service cost using the Google Cloud pricing calculator and billing reports and by analyzing billing data. Observe whether your services are meeting their SLOs using Cloud Monitoring and Dashboards. Use Uptime Checks to determine service availability. Respond to service outages using Cloud Monitoring Alerts. Additional course details: Nexus Humans Architecting with Google Cloud: Design and Process training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Architecting with Google Cloud: Design and Process course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.