Duration 2 Days 12 CPD hours This course is intended for This course is ideal for beginners and intermediate frontend developers who want to become full-stack developers. You will need some prior working knowledge of JavaScript and MongoDB as we skim over its basics and get straight to work. Overview At the end of this day, students should be able to: Understand the MEAN architecture Create RESTful APIs to complete CRUD tasks Build a blogging application with basic features Describe best practices to secure node applications Implement authentication and authorization Create simple animations using Angular Perform unit testing on Angular applications MongoDB, Express, Angular and Node.js Fundamentals begins by demystifying the MEAN architecture. You will review the features of the JavaScript technologies, frameworks, or libraries that make up a MEAN stack. You will also understand how to develop a RESTful API using Node.js, Express.js, and MongoDB Atlas. This course will enable you to discover how to build a blogging application using the MEAN stack. Next, you will learn about authentication using MEAN, and explore the features of Angular, such as pipes, reactive forms, modules and optimizing apps, animations and unit testing, and much more. By the end of the course, you will have all of the knowledge you need to become a pro at developing efficient web applications using JavaScript technologies. Introduction to the MEAN stack MEAN Architecture Demystification Getting Started with Node Activity 1: Creating an HTTP Server for a Blogging Application Understanding Callbacks, Event loop and EventEmitters in Node Understanding Buffers, Streams and Filesystem in Node Activity 2: Streaming Data to a File Developing RESTful APIs to perform CRUD operations Getting Started with RESTful APIs Getting started with MongoDB Atlas Activity 3: Connecting the Node Application with MongoDB Atlas Getting Started with Express Activity 4: Creating Express API Route and Controller Activity 5: Testing Fully Functional RESTful API Beginning Frontend Development with Angular CLI Getting Started with Angular CLI Using Components, Directives, Services, and Making HTTP Requests in Angular Activity 6: Designing the Frontend and Components for the Blogging Application Activity 7: Writing Services and Making HTTP Request Calls to an API Understanding Angular Forms and Routing Activity 8: Creating a Form Application Using the Reactive/Model-Driven Method Activity 9: Creating and Validating Different Forms Using the Template and Reactive Driven Method Activity 10: Implementing a Router for the Blogging Application Understanding MEAN Stack Security Node Security and Best Practices Node Application Authentication with JSON Web Token (JWT) Activity 11: Securing the RESTful API Node Application Authentication with Passport Activity 12: Creating a Login Page to Allow Authentication with Twitter Using Passport Strategies Angular Declarables, Bootstrapping, and Modularity Using Inbuilt Pipes, Custom Pipes, Custom Directives, and Observables Activity 13: Communicating Between Two Components Using Observable Angular Bootstrapping and Modularity Activity 14: Creating a Lazy Loaded Application Testing and Optimizing Angular Applications Angular Animations and Latest Angular Features Activity 15: Animating the Route Transition Between the Blog Post Page and View Post Page of the Blogging Application Optimizing Angular Applications Testing Angular Applications Activity 16: Performing Unit Testing on the App Root Component and Blog-Post Component Overview on the new features in Angular
Duration 5 Days 30 CPD hours This course is intended for Linux system administrators, site reliability engineers, and other IT professionals with some Ansible experience who are interested in learning how to manage and automate the deployment, configuration, and operation of key network services included with Red Hat Enterprise Linux 8. Overview Provide key network services using software included with Red Hat Enterprise Linux 8, including DNS with Unbound and BIND9, DHCP and DHCPv6, client e-mail transmission, printing service, NFS and SMB protocol file sharing, SQL database service with MariaDB, and web services using Apache HTTPD, nginx, Varnish, and HAProxy. Configure advanced networking for server use cases, including device teaming. Use Red Hat Ansible Engine to automate the manual deployment and configuration tasks covered in this course. Learn how to configure, manage, and scale key services used in the data center Red Hat Services Management and Automation (RH358) is designed for IT professionals with some experience managing Linux© systems and want to learn more about how to manage and deploy network services included with Red Hat© Enterprise Linux which are particularly important in the modern IT data center. You will learn how to install, configure, and manage basic configurations of these services manually, and then use Red Hat Ansible© Engine to automate your work in a scalable, repeatable manner. This course is based on Red Hat Ansible Engine 2.9 and Red Hat Enterprise Linux 8.1. 1 - Manage network services Discuss and review key tools and skills needed to manage network services. 2 - Configure link aggregation Improve the redundancy or throughput of network connections of servers by configuring Linux network teaming between multiple network interfaces. 3 - Manage DNS and DNS servers Explain the operation of DNS service, troubleshoot DNS issues, and configure servers to act as a DNS caching nameserver or as an authoritative name server. 4 - Manage DHCP and IP address assignment Explain and configure services used for IPv4 and IPv6 address assignment including DHCP, DHCPv6, and SLAAC. 5 - Manage printers and printing files Configure systems to print to a network printer that supports IPP Everywhere, as well as manage existing printer queues. 6 - Configure email transmission Discuss how mail servers operate, then configure a server to use system tools and Postfix to send email messages through an outbound mail relay. 7 - Configure MariaDB SQL databases Discuss the basic operation of SQL-based relational databases, perform basic SQL queries for troubleshooting, and be able to set up a simple MariaDB database service. 8 - Configure web servers Provide web content from Apache HTTPD or Nginx web servers, then configure them with virtual hosts and TLS-based encryption. 9 - Optimize web server traffic Improve performance of your web servers by using Varnish to cache static content being served and HAProxy to terminate TLS connections and balance load between servers. 10 - Provide file-based network storage Deliver simple file-based network shares to clients using the NFS and SMB protocols. 11 - Access block-based network storage Configure iSCSI initiators on your servers to access block-based storage devices provided by network storage arrays or Ceph storage clusters.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Architects and operators who build and manage data analytics pipelines Overview In this course, you will learn to: Compare the features and benefits of data warehouses, data lakes, and modern data architectures Design and implement a batch data analytics solution Identify and apply appropriate techniques, including compression, to optimize data storage Select and deploy appropriate options to ingest, transform, and store data Choose the appropriate instance and node types, clusters, auto scaling, and network topology for a particular business use case Understand how data storage and processing affect the analysis and visualization mechanisms needed to gain actionable business insights Secure data at rest and in transit Monitor analytics workloads to identify and remediate problems Apply cost management best practices In this course, you will learn to build batch data analytics solutions using Amazon EMR, an enterprise-grade Apache Spark and Apache Hadoop managed service. You will learn how Amazon EMR integrates with open-source projects such as Apache Hive, Hue, and HBase, and with AWS services such as AWS Glue and AWS Lake Formation. The course addresses data collection, ingestion, cataloging, storage, and processing components in the context of Spark and Hadoop. You will learn to use EMR Notebooks to support both analytics and machine learning workloads. You will also learn to apply security, performance, and cost management best practices to the operation of Amazon EMR. Module A: Overview of Data Analytics and the Data Pipeline Data analytics use cases Using the data pipeline for analytics Module 1: Introduction to Amazon EMR Using Amazon EMR in analytics solutions Amazon EMR cluster architecture Interactive Demo 1: Launching an Amazon EMR cluster Cost management strategies Module 2: Data Analytics Pipeline Using Amazon EMR: Ingestion and Storage Storage optimization with Amazon EMR Data ingestion techniques Module 3: High-Performance Batch Data Analytics Using Apache Spark on Amazon EMR Apache Spark on Amazon EMR use cases Why Apache Spark on Amazon EMR Spark concepts Interactive Demo 2: Connect to an EMR cluster and perform Scala commands using the Spark shell Transformation, processing, and analytics Using notebooks with Amazon EMR Practice Lab 1: Low-latency data analytics using Apache Spark on Amazon EMR Module 4: Processing and Analyzing Batch Data with Amazon EMR and Apache Hive Using Amazon EMR with Hive to process batch data Transformation, processing, and analytics Practice Lab 2: Batch data processing using Amazon EMR with Hive Introduction to Apache HBase on Amazon EMR Module 5: Serverless Data Processing Serverless data processing, transformation, and analytics Using AWS Glue with Amazon EMR workloads Practice Lab 3: Orchestrate data processing in Spark using AWS Step Functions Module 6: Security and Monitoring of Amazon EMR Clusters Securing EMR clusters Interactive Demo 3: Client-side encryption with EMRFS Monitoring and troubleshooting Amazon EMR clusters Demo: Reviewing Apache Spark cluster history Module 7: Designing Batch Data Analytics Solutions Batch data analytics use cases Activity: Designing a batch data analytics workflow Module B: Developing Modern Data Architectures on AWS Modern data architectures
Duration 3 Days 18 CPD hours This course is intended for This course is appropriate for anyone needing to interface with an Oracle database or those needing a general understanding of Oracle database functionality. That would include end users, business analysts, application developers and database administrators. Overview Working in a hands-on learning environment led by our expert pracitioner you'll learn how to: Add Data, Retrieve, Sort and Organize a SQL Database Combine Data, Set Operators and Subqueries Manipulate Data and Data Definition Languages in SQL Work with Data Dictionary Views and Create Sequences, Indexes and Views Use Database Objects and Subqueries Perform Data and access control Perform other Advanced Level Database operations. Oracle 19C SQL Programming Fundamentals Is a three-day, hands-on course designed to equip you with the fundamental skills needed to set up, run and manage SQL databases using Oracle Database Technology. You will also be discovering all the tools and concepts required to organize data efficiently. Our engaging instructors and mentors are highly-experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Working within in a hands-on learning environment, guided by our expert team, attendees will develop a practical approach to Oracle Database Technology. Throughout the course, you will learn the key elements of a database, and the way Oracle systems facilitate their induction in the system. You?ll also learn the tools and strategies you can implement to store, retrieve, compare and organize data according to your requirements. You?ll also explore the process of creating simple to complex reports from existing data. By the end of this course, you will also have hands-on knowledge of SQL systems that are required to proceed to other advanced to professional programs. Adding Data, Retrieving, Sorting and Organizing a SQL Database The building blocks of a database. How to add data to the database. The process of retrieving data using SQL functions. Multiple methods of sorting and organizing data. Using SQL functions to get the required simple to complex output. Various strategies for using functions and conditions to organize data. Combining Data, Set Operators And Subqueries Consolidating data using multiple functions and group operators. Fetching intelligent data reports using simple functions. Fetching data from multiple sources in the tables. Using Subqueries to compile data as required. Using Set operators to create smart data reports. Data Manipulation and Data Definition Languages in SQL Describing and managing data transaction using Data Definition Language. Categorize and review data tables using Data Definition Language. Data Dictionary Views and Creating Sequences, Indexes and Views How to manage and query Data Dictionary Views. The process of creating and using Sequences. How to create various types of Synonyms and Indexes. Creating simple and complex views and retrieving data. Understanding and using Database Objects and Subqueries Core concept and application of Schema Objects. Fetching required data with Subqueries. Using Subqueries to organize Data in SQL. Data and access control Assigning and revoking data access. Managing data access control according to user levels. Performing Advanced Level Database operations. Using advanced functions and performing data queries. Creating and managing time zone-based databases.
Duration 4 Days 24 CPD hours This course is intended for This course benefits individuals responsible for configuring and monitoring EX Series switches running Junos ELS. Overview After successfully completing this course, you should be able to: List the benefits of implementing switched LANs. Describe transparent bridging concepts and operations. Describe terms and design considerations for switched LANs. List enterprise platforms that support Layer 2 switching. Configure interfaces for Layer 2 switching operations. Display and interpret the Ethernet switching table. Explain the concept of a VLAN. Describe access and trunk port modes. Configure and monitor VLANs. Describe voice VLAN and native VLAN concepts. Explain inter-VLAN routing operations. Configure and monitor inter-VLAN routing. Explain when a spanning tree is required. Describe STP and Rapid Spanning Tree Protocol (RSTP) operations. List some advantages of using RSTP over STP. Configure and monitor RSTP. Describe the bridge protocol data unit (BPDU), loop, and root protection features. Configure and monitor the BPDU, loop, and root protection features. List and describe various port security features. Configure and monitor port security features. Describe the storm control feature. Configure and monitor storm control. Describe firewall filter support for EX Series Ethernet Switches. Implement and monitor the effects of a firewall filter. List and describe some features that promote high availability. Configure and monitor high availability features. Describe the basic concepts and operational details of a virtual chassis. Implement a virtual chassis with multiple EX4300 switches. Explain the concepts of Ethernet Ring Protection Switching (ERPS). Configure and monitor ERPS. Explain the concepts of Multiple Spanning Tree Protocol (MSTP). Configure and monitor MSTP. This 4-day program includes an overview of protocol independent routing features, load balancing and filter-based forwarding, OSPF, BGP, IP tunneling, switching concepts and operations, virtual LANs (VLANs), the Spanning Tree Protocol (STP), and more. Chapter 1: Course Introduction Course Introduction Chapter 2: Layer 2 Switching Ethernet Bridging Basics Terminology and Design Considerations Overview of Enterprise Switching Platforms Enabling and Monitoring Layer 2 Switching Operations Lab 1: Implementing Layer 2 Switching Chapter 3: Virtual Networks Overview of VLANs Configuring and Monitoring VLANs Voice VLAN Native VLAN Routed VLAN Interfaces Lab 2: Implementing Virtual Networks Chapter 4: Routing Instances Routing Instances Overview Configuring and Monitoring Virtual Switches Chapter 5: Spanning Tree Spanning Tree Protocol Rapid Spanning Tree Protocol Configuring and Monitoring STP and RSTP Protection Features: BPDU Protection Protection Features: Loop Protection Protection Features: Root Protection Lab 3: Implementing Spanning Tree Chapter 6: Port Security MAC Limiting Persistent MAC Learning DHCP Snooping Dynamic ARP Inspection (DAI) IP Source Guard Lab 4: Implementing Port Security Chapter 7: Device Security and Firewall Filters Storm Control Firewall Filters Lab 5: Implementing Storm Control and Firewall Filters Chapter 8: Virtual Chassis Overview of Virtual Chassis Configuring and Monitoring a Virtual Chassis Lab 6: Implementing a Virtual Chassis System Chapter 9: High Availability Features Overview of High Availability Networks Link Aggregation Groups Redundant Trunk Groups Graceful Routing Engine Switchover (GRES) Nonstop Active Routing (NSR) Nonstop Bridging (NSB) Lab 7: Implementing High Availability Features
Duration 2 Days 12 CPD hours This course is intended for This course is intended for Data Warehouse Administrator, Database Administrator, Database Designer, Support Engineer, and Technical Administrator. Overview Upon successful completion of this course, students will be able to Install Oracle Grid Infrastructure for a Standalone Server, Configure ASMLIB and create ASM disk groups, Apply the latest patch set updates to the Grid Infrastructure software, Install Oracle Database software, Use Oracle Restart to manage components, Prepare the Linux operating system for Oracle software installation, Upgrade Oracle Database, Create a container database, Install Oracle Grid Infrastructure for an Independent Server, Create an Oracle Database using DBCA, and Install the latest patch set updates to the Oracle database software. This Oracle Database: Deploy, Patch and Upgrade Workshop gives you detailed information to help you install Oracle Database software. In this course, you will be introduced to Oracle Database Cloud Service. Oracle Database Overview Oracle Database Introduction Oracle Database Architecture Overview Oracle Database Instance Configurations Oracle Database Memory Structures Process Structures Database Storage Architecture Logical and Physical Database Structures Container and Pluggable Database Overview Installing Oracle Grid Infrastructure for a Standalone Server Overview of Oracle Grid Infrastructure for a Standalone Server System Requirements for Oracle Grid Infrastructure Configuring Storage for Oracle Automatic Storage Management (ASM) Installing Oracle Grid Infrastructure for a Standalone Server Upgrading Oracle Grid Infrastructure for a Standalone Server Installing Oracle Database Software Planning Your Installation System Requirements for Oracle Database Preparing the Operating System Using 4 KB Sector Disks Setting Environment Variables Checking the System Requirements Using the Oracle Universal Installer (OUI) Performing a Silent Mode Installation Creating an Oracle Database by Using DBCA Planning the Database Storage Structure Choosing non-CDB or CDB Types of Databases (based on workload) Choosing the Appropriate Character Set Understanding How Character Sets are Used Setting the NLS_LANG Initialization Parameter Using the Database Configuration Assistant (DBCA) Using Oracle Restart Oracle Restart Overview Oracle Restart Process startup Controlling Oracle Restart Choosing the Correct SRVCTL Utility Oracle Restart Configuration Using the SRVCTL Utility Obtaining Help for the SRVCTL Utility Starting Components by Using the SRVCTL Utility Introduction to Upgrading to Oracle Database Upgrade Methods Data Migration Methods Supported Releases for Direct Upgrade Overview of Upgrade Process Performing a Rolling Upgrade Upgrading a CBD Preparing to Upgrade to Oracle Database Developing a Test Plan Performance Testing Requirements for Databases Using Oracle Label Security or Oracle Database Vault Requirement for Databases Using Oracle Warehouse Builder Using the Pre-Upgrade Information Tool Backing Up the Database Installing the Oracle Database Software Preparing the New Oracle Home Upgrading to Oracle Database Upgrading by Using the Database Upgrade Assistant (DBUA) Manually Upgrading to Oracle Database Migrating a non-CDB to a CDB Performing Post-Upgrade Tasks Migrating to Unified Auditing Performing Post-Upgrade Tasks Following a Manual Upgrade Migrating Data by Using Oracle Data Pump Data Pump Overview Migrating by Using Data Pump Importing by Using a Network Link
Duration 4 Days 24 CPD hours This course is intended for This course is best suited to developers, engineers, and architects who want to use use Hadoop and related tools to solve real-world problems. Overview Skills learned in this course include:Creating a data set with Kite SDKDeveloping custom Flume components for data ingestionManaging a multi-stage workflow with OozieAnalyzing data with CrunchWriting user-defined functions for Hive and ImpalaWriting user-defined functions for Hive and ImpalaIndexing data with Cloudera Search Cloudera University?s four-day course for designing and building Big Data applications prepares you to analyze and solve real-world problems using Apache Hadoop and associated tools in the enterprise data hub (EDH). IntroductionApplication Architecture Scenario Explanation Understanding the Development Environment Identifying and Collecting Input Data Selecting Tools for Data Processing and Analysis Presenting Results to the Use Defining & Using Datasets Metadata Management What is Apache Avro? Avro Schemas Avro Schema Evolution Selecting a File Format Performance Considerations Using the Kite SDK Data Module What is the Kite SDK? Fundamental Data Module Concepts Creating New Data Sets Using the Kite SDK Loading, Accessing, and Deleting a Data Set Importing Relational Data with Apache Sqoop What is Apache Sqoop? Basic Imports Limiting Results Improving Sqoop?s Performance Sqoop 2 Capturing Data with Apache Flume What is Apache Flume? Basic Flume Architecture Flume Sources Flume Sinks Flume Configuration Logging Application Events to Hadoop Developing Custom Flume Components Flume Data Flow and Common Extension Points Custom Flume Sources Developing a Flume Pollable Source Developing a Flume Event-Driven Source Custom Flume Interceptors Developing a Header-Modifying Flume Interceptor Developing a Filtering Flume Interceptor Writing Avro Objects with a Custom Flume Interceptor Managing Workflows with Apache Oozie The Need for Workflow Management What is Apache Oozie? Defining an Oozie Workflow Validation, Packaging, and Deployment Running and Tracking Workflows Using the CLI Hue UI for Oozie Processing Data Pipelines with Apache Crunch What is Apache Crunch? Understanding the Crunch Pipeline Comparing Crunch to Java MapReduce Working with Crunch Projects Reading and Writing Data in Crunch Data Collection API Functions Utility Classes in the Crunch API Working with Tables in Apache Hive What is Apache Hive? Accessing Hive Basic Query Syntax Creating and Populating Hive Tables How Hive Reads Data Using the RegexSerDe in Hive Developing User-Defined Functions What are User-Defined Functions? Implementing a User-Defined Function Deploying Custom Libraries in Hive Registering a User-Defined Function in Hive Executing Interactive Queries with Impala What is Impala? Comparing Hive to Impala Running Queries in Impala Support for User-Defined Functions Data and Metadata Management Understanding Cloudera Search What is Cloudera Search? Search Architecture Supported Document Formats Indexing Data with Cloudera Search Collection and Schema Management Morphlines Indexing Data in Batch Mode Indexing Data in Near Real Time Presenting Results to Users Solr Query Syntax Building a Search UI with Hue Accessing Impala through JDBC Powering a Custom Web Application with Impala and Search
Duration 2 Days 12 CPD hours This course is intended for Report Authors Overview Create query models Create reports based on query relationships Introduction to dimensional data Introduction to dimensional data in reports Dimensional report context Focus your dimensional data Calculations and dimensional functions Create advanced dynamic reports This offering teaches Professional Report Authors about advanced report building techniques using relational data models, dimensional data, and ways of enhancing, customizing, managing, and distributing professional reports. The course builds on topics presented in the Fundamentals course. Activities will illustrate and reinforce key concepts during this learning activity. Create query models Build a query and connect it to a report Answer a business question by referencing data in a separate query Create reports based on query relationships Create join relationships between queries Combine data containers based on relationships from different queries Create a report comparing the percentage of change Introduction to dimensional reporting concepts Examine data sources and model types Describe the dimensional approach to queries Apply report authoring styles Introduction to dimensional data in reports Use members to create reports Identify sets and tuples in reports Use query calculations and set definitions Dimensional report context Examine dimensional report members Examine dimensional report measures Use the default measure to create a summarized column in a report Focus your dimensional data Focus your report by excluding members of a defined set Compare the use of the filter() function to a detail filter Filter dimensional data using slicers Calculations and dimensional functions Examine dimensional functions Show totals and exclude members Create a percent of base calculation Create advanced dynamic reports Use query macros Control report output using a query macro Create a dynamic growth report Create a report that displays summary data before detailed data and uses singletons to summarize information Design effective prompts Create a prompt that allows users to select conditional formatting values Create a prompt that provides users a choice between different filters Create a prompt to let users choose a column sort order Create a prompt to let users select a display type Examine the report specification Examine report specification flow Identify considerations when modifying report specifications Customize reporting objects Distribute reports Burst a report to email recipients by using a data item Burst a list report to the IBM Cognos Analytics portal by using a burst table Burst a crosstab report to the IBM Cognos Analytics portal by using a burst table and a master detail relationship Enhance user interaction with HTML Create interactive reports using HTML Include additional information with tooltips Send emails using links in a report Introduction to IBM Cognos Active Reports Examine Active Report controls and variables Create a simple Active Report using Static and Data-driven controls Change filtering and selection behavior in a report Create interaction between multiple controls and variables Active Report charts and decks Create an Active Report with a Data deck Use Master detail relationships with Decks Optimize Active Reports Create an Active Report with new visualizations
Duration 3.75 Days 22.5 CPD hours This course is intended for New systems administrators Overview When you complete this course, you will be able to:Customize your application, including page layouts, fields, tabs, and business processes in Lightning Experience.Learn how security settings created in Salesforce Classic are applied in Lightning.Maintain and import clean data in Lightning.Use Lightning features to create high-value reports and dashboards.Understand how workflow automation complies with Lightning This course is the core training that ensures your success with Salesforce Lightning. It?s a must for new administrators, and we recommend completing this course before starting a Salesforce deployment or when taking over an existing deployment. Getting Around the App Data Model and Navigation Lightning Experience Help & Training Getting Your Organization Ready for Users Setting Up the Company Profile Configuring the User Interface Setting Up Activities and Calendars Configuring Search Settings Setting Up Chatter Groups Mobile Access with Salesforce1 Setting Up & Managing Users Managing User Profiles Managing Users Setting Up Chatter Free Users and Invites Troubleshooting Login Issues Security & Data Access Restricting Logins Determining Object Access Setting Up Record Access Creating a Role Hierarchy Dealing with Record Access Exceptions Managing Field-level Security Object Customizations Administering Standard Fields Creating New Custom Fields Creating Selection Fields: Picklists and Lookups Creating Formula Fields Working with Page Layouts Working with Record Types and Business Processes Maintaining Data Quality Managing Data Import Wizards Data Loader Data.com Mass Transfer Backing Up Data Mass Delete and Recycle Bin Reports & Dashboards Running and Modifying Reports Creating New Reports with the Report Builder Working with Report Filters Summarizing with Formulas and Visual Summaries Printing, Exporting, and Emailing Reports Building Dashboards Automation Email Templates Workflow Rules Process Builder Lead Automation Managing the Support Process Managing and Resolving Cases Customizing a Support Process Automating Support Understanding the Salesforce Console for Service Collaborating in the Service Cloud Analyzing Support Data Additional course details: Nexus Humans Salesforce Administration Essentials for New Admins in Lightning Experience (ADX201) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Salesforce Administration Essentials for New Admins in Lightning Experience (ADX201) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 3 Days 18 CPD hours This course is intended for This course is intended for: Database architects Database administrators Database developers Data analysts and scientists Overview This course is designed to teach you how to: Discuss the core concepts of data warehousing, and the intersection between data warehousing and big data solutions Launch an Amazon Redshift cluster and use the components, features, and functionality to implement a data warehouse in the cloud Use other AWS data and analytic services, such as Amazon DynamoDB, Amazon EMR, Amazon Kinesis, and Amazon S3, to contribute to the data warehousing solution Architect the data warehouse Identify performance issues, optimize queries, and tune the database for better performance Use Amazon Redshift Spectrum to analyze data directly from an Amazon S3 bucket Use Amazon QuickSight to perform data analysis and visualization tasks against the data warehouse Data Warehousing on AWS introduces you to concepts, strategies, and best practices for designing a cloud-based data warehousing solution using Amazon Redshift, the petabyte-scale data warehouse in AWS. This course demonstrates how to collect, store, and prepare data for the data warehouse by using other AWS services such as Amazon DynamoDB, Amazon EMR, Amazon Kinesis, and Amazon S3. Additionally, this course demonstrates how to use Amazon QuickSight to perform analysis on your data Module 1: Introduction to Data Warehousing Relational databases Data warehousing concepts The intersection of data warehousing and big data Overview of data management in AWS Hands-on lab 1: Introduction to Amazon Redshift Module 2: Introduction to Amazon Redshift Conceptual overview Real-world use cases Hands-on lab 2: Launching an Amazon Redshift cluster Module 3: Launching clusters Building the cluster Connecting to the cluster Controlling access Database security Load data Hands-on lab 3: Optimizing database schemas Module 4: Designing the database schema Schemas and data types Columnar compression Data distribution styles Data sorting methods Module 5: Identifying data sources Data sources overview Amazon S3 Amazon DynamoDB Amazon EMR Amazon Kinesis Data Firehose AWS Lambda Database Loader for Amazon Redshift Hands-on lab 4: Loading real-time data into an Amazon Redshift database Module 6: Loading data Preparing Data Loading data using COPY Data Warehousing on AWS AWS Classroom Training Concurrent write operations Troubleshooting load issues Hands-on lab 5: Loading data with the COPY command Module 7: Writing queries and tuning for performance Amazon Redshift SQL User-Defined Functions (UDFs) Factors that affect query performance The EXPLAIN command and query plans Workload Management (WLM) Hands-on lab 6: Configuring workload management Module 8: Amazon Redshift Spectrum Amazon Redshift Spectrum Configuring data for Amazon Redshift Spectrum Amazon Redshift Spectrum Queries Hands-on lab 7: Using Amazon Redshift Spectrum Module 9: Maintaining clusters Audit logging Performance monitoring Events and notifications Lab 8: Auditing and monitoring clusters Resizing clusters Backing up and restoring clusters Resource tagging and limits and constraints Hands-on lab 9: Backing up, restoring and resizing clusters Module 10: Analyzing and visualizing data Power of visualizations Building dashboards Amazon QuickSight editions and feature