Duration 5 Days 30 CPD hours This course is intended for This course is intended for both novice and experienced programmers. Some experience in C++, C# or Java is helpful. Overview After completing this course, you will be able to: Write and test JavaScript code in modern browsers. Understand and work with the browser's Document Object Model (DOM). Work DOM objects and write code to respond to events. Add jQuery to their projects and use jQuery to manipulate the DOM and to make Ajax calls. Understand Ajax and Web Services, including HTTP Verbs and asynchronous Ajax calls. Work with XML and JSON data formats. Understand and use RESTful Web Services including ODATA queries. Work with Anonymous Functions, Self-Invoking Functions and Closures. Understand the role of the Window and Navigator objects. Describe and understand variables, 'variable hoisting', and JavaScript objects. This course is an in depth hands-on study of JavaScript. The course includes detailed hands-on labs and Q&A labs. The labs include multiple projects that update an existing web site. This course is similar to the retired Course 20480: Programming in HTML5 with JavaScript and CSS3. Module 1: Basic Scripting JavaScript Basics JavaScript or ECMAScript? Which Version of JavaScript am I Using? Shims, Shivs, and Polyfills Writing and Testing JavaScript Editors and the F12 Tools Key Parts of a Script Module 2: The Document Object Model The DOM Accessing Objects from the DOM Responding to Events Adding Elements to the DOM Module 3: Introduction to jQuery jQuery Downloading the jQuery Library jQuery Versions Linking to the Library Selecting Elements using jQuery Working with the Data Returned by jQuery Setting CSS Properties Running a Function Against the jQuery Return Set Module 4: Ajax and Web Services Web Services HTTP Verbs Data Formats ? XML and JSON Ajax RESTful Web Services OData OData Queries OData Updates Module 5: Diving in Deeper Is JavaScript Compiled? Data Types and Objects Extending Data Types Variables, Scopes and Hoisting Functions The ?this? Keyword Errors Recursion
Duration 3 Days 18 CPD hours This course is intended for Data Analysts, Business Analysts, Business Intelligence professionals Cloud Data Engineers who will be partnering with Data Analysts to build scalable data solutions on Google Cloud Platform Overview This course teaches students the following skills: Derive insights from data using the analysis and visualization tools on Google Cloud Platform Interactively query datasets using Google BigQuery Load, clean, and transform data at scale Visualize data using Google Data Studio and other third-party platforms Distinguish between exploratory and explanatory analytics and when to use each approach Explore new datasets and uncover hidden insights quickly and effectively Optimizing data models and queries for price and performance Want to know how to query and process petabytes of data in seconds? Curious about data analysis that scales automatically as your data grows? Welcome to the Data Insights course! This four-course accelerated online specialization teaches course participants how to derive insights through data analysis and visualization using the Google Cloud Platform. The courses feature interactive scenarios and hands-on labs where participants explore, mine, load, visualize, and extract insights from diverse Google BigQuery datasets. The courses also cover data loading, querying, schema modeling, optimizing performance, query pricing, and data visualization. This specialization is intended for the following participants: Data Analysts, Business Analysts, Business Intelligence professionals Cloud Data Engineers who will be partnering with Data Analysts to build scalable data solutions on Google Cloud Platform To get the most out of this specialization, we recommend participants have some proficiency with ANSI SQL. Introduction to Data on the Google Cloud Platform Highlight Analytics Challenges Faced by Data Analysts Compare Big Data On-Premises vs on the Cloud Learn from Real-World Use Cases of Companies Transformed through Analytics on the Cloud Navigate Google Cloud Platform Project Basics Lab: Getting started with Google Cloud Platform Big Data Tools Overview Walkthrough Data Analyst Tasks, Challenges, and Introduce Google Cloud Platform Data Tools Demo: Analyze 10 Billion Records with Google BigQuery Explore 9 Fundamental Google BigQuery Features Compare GCP Tools for Analysts, Data Scientists, and Data Engineers Lab: Exploring Datasets with Google BigQuery Exploring your Data with SQL Compare Common Data Exploration Techniques Learn How to Code High Quality Standard SQL Explore Google BigQuery Public Datasets Visualization Preview: Google Data Studio Lab: Troubleshoot Common SQL Errors Google BigQuery Pricing Walkthrough of a BigQuery Job Calculate BigQuery Pricing: Storage, Querying, and Streaming Costs Optimize Queries for Cost Lab: Calculate Google BigQuery Pricing Cleaning and Transforming your Data Examine the 5 Principles of Dataset Integrity Characterize Dataset Shape and Skew Clean and Transform Data using SQL Clean and Transform Data using a new UI: Introducing Cloud Dataprep Lab: Explore and Shape Data with Cloud Dataprep Storing and Exporting Data Compare Permanent vs Temporary Tables Save and Export Query Results Performance Preview: Query Cache Lab: Creating new Permanent Tables Ingesting New Datasets into Google BigQuery Query from External Data Sources Avoid Data Ingesting Pitfalls Ingest New Data into Permanent Tables Discuss Streaming Inserts Lab: Ingesting and Querying New Datasets Data Visualization Overview of Data Visualization Principles Exploratory vs Explanatory Analysis Approaches Demo: Google Data Studio UI Connect Google Data Studio to Google BigQuery Lab: Exploring a Dataset in Google Data Studio Joining and Merging Datasets Merge Historical Data Tables with UNION Introduce Table Wildcards for Easy Merges Review Data Schemas: Linking Data Across Multiple Tables Walkthrough JOIN Examples and Pitfalls Lab: Join and Union Data from Multiple Tables Advanced Functions and Clauses Review SQL Case Statements Introduce Analytical Window Functions Safeguard Data with One-Way Field Encryption Discuss Effective Sub-query and CTE design Compare SQL and Javascript UDFs Lab: Deriving Insights with Advanced SQL Functions Schema Design and Nested Data Structures Compare Google BigQuery vs Traditional RDBMS Data Architecture Normalization vs Denormalization: Performance Tradeoffs Schema Review: The Good, The Bad, and The Ugly Arrays and Nested Data in Google BigQuery Lab: Querying Nested and Repeated Data More Visualization with Google Data Studio Create Case Statements and Calculated Fields Avoid Performance Pitfalls with Cache considerations Share Dashboards and Discuss Data Access considerations Optimizing for Performance Avoid Google BigQuery Performance Pitfalls Prevent Hotspots in your Data Diagnose Performance Issues with the Query Explanation map Lab: Optimizing and Troubleshooting Query Performance Advanced Insights Introducing Cloud Datalab Cloud Datalab Notebooks and Cells Benefits of Cloud Datalab Data Access Compare IAM and BigQuery Dataset Roles Avoid Access Pitfalls Review Members, Roles, Organizations, Account Administration, and Service Accounts
Duration 5 Days 30 CPD hours This course is intended for This course is intended for: Solutions Architects who are new to designing and building cloud architectures Data Center Architects who are migrating from on-premises environment to cloud architectures Other IT/cloud roles who want to understand how to design and build cloud architectures Overview In this course, you will learn how to: Make architectural decisions based on AWS architectural principles and best practices Use AWS services to make your infrastructure scalable, reliable, and highly available Use AWS Managed Services to enable greater flexibility and resiliency in an infrastructure Make an AWS-based infrastructure more efficient to increase performance and reduce costs Use the Well Architected Framework to improve architectures with AWS solutions This course covers all aspects of how to architect for the cloud over four-and-a-half-days. It covers topics from Architecting on AWS and Advanced Architecting on AWS to offer an immersive course in cloud architecture. You will learn how to design cloud architectures, starting small and working to large-scale enterprise level designs-and everything in between. Starting with the Well-Architected Framework, you will learn important architecting information for AWS services including: compute, storage, database, networking, security, monitoring, automation, optimization, benefits of de-coupling applications and serverless, building for resilience, and understanding costs Module 1: Introduction The real story of AWS Well-Architected Framework Six advantages of the cloud Global infrastructure Module 2: The Simplest Architectures S3 Glacier Choosing your regions Hands-on lab: Static Website Module 3: Adding a Compute Layer EC2 Storage solutions for instances Purchasing options such as dedicated host vs instances Module 4: Adding a Database Layer Relational vs non-relational Managed databases RDS Dynamo DB Neptune Hands-on lab: Deploying a web application on AWS Module 5: Networking in AWS Part 1 VPC CIDR and subnets Public vs private subnets NAT and internet gateway Security groups Module 6: Networking in AWS Part 2 Virtual Private Gateway VPN Direct Connect VPC peering Transit Gateway VPC Endpoints Elastic Load Balancer Route 53 Hands-on lab: Creating a VPC Module 7: AWS Identity and Access Management (IAM) IAM Identity federation Cognito Module 8: Organizations Organizations Multiple account management Tagging strategies Module 9: Elasticity, High Availability, and Monitoring Elasticity vs inelasticity Monitoring with CloudWatch, CloudTrail, and VPC Flow Logs Auto scaling Scaling databases Hands-on lab: Creating a highly available environment Module 10: Automation Why automate? CloudFormation AWS Quick Starts AWS Systems Manager AWS OpsWorks AWS Elastic Beanstalk Module 11: Deployment Methods Why use a deployment method? Blue green and canary deployment Tools to implement your deployment methods CI/CD Hands-on lab: Automating infrastructure deployment Module 12: Caching When and why you should cache your data Cloudfront Elasticache (Redis/Memcached) DynamoDB Accelerator Module 13: Security of Your Data Shared responsibility model Data classification Encryption Automatic data security Module 14: Building Decoupled Architecture Tight coupling vs loose coupling SQS SNS Module 15: Optimizations and Review Review questions Best practices Activity: Design and architecture - two trues and one lie Module 16: Microservices What is a microservice? Containers ECS Fargate EKS Module 17: Serverless Why use serverless? Lambda API Gateway AWS Step Functions Hands-on lab: Implementing a serverless architecture with AWS Managed Services Module 18: Building for Resilience Using managed services greatly increases resiliency Serverless for resiliency Issues with microservices to be aware of DDoS Hands-on lab: Amazon CloudFront content delivery and automating WAF rules Module 19: Networking in AWS Part 3 Elastic Network Adapter Maximum transmission units Global Accelerator Site to site VPN Transit Gateway Module 20: Understanding Costs Simple monthly calculator Right sizing your instances Price sensitive architecture examples Module 21: Migration Strategies Cloud migration strategies Planning Migrating Optimizing Hands-on lab: Application deployment using AWS Fargate Module 22: RTO/RPO and Backup Recovery Setup Disaster planning Recovery options Module 23: Final Review Architecting advice Service use case questions Example test questions Additional course details: Nexus Humans Architecting on AWS - Accelerator training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Architecting on AWS - Accelerator course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 4 Days 24 CPD hours This course is intended for This class is designed for enterprise architects, solution architects, and business analysts working to earn their Salesforce Application Architect credential, or for application architects looking to get more hands-on experience. Overview Design data models that scale gracefully Leverage Salesforce sharing mechanisms at an advanced level Evaluate the nuances of field types and pick the right one for the circumstances Make data model decisions that minimize record locking and other performance degradations Dive into the two cornerstone domains of being an Application Architect: Data Architecture and Management, and Sharing and Visibility. In this 4-day workshop, our Architect experts will present you with a case study scenario that will be broken down and digested through iterative exploration. Learn how to design and build secure, scalable, and high-performing solutions through a combination of lecture, demos, hands-on exercises, and workshop presentations and discussions. Application Architect Overview Assess the Requirements to Become an Application Architect Understand the Real-World Expectations of Application Architects Review the Core Characteristics of Successful Application Architects Scenario Introduction Review the Application Architecture Scenario Identify Scenario Problem Areas Identify Scenario Actors and Licensing Architecture Documentation Understand Key Architecture Documentation Requirements Learn About Best Practices for Artifact Documentation Produce Architecture Documents Data Modeling Identify Relationship Types and Their Impact on Record Access, User Interface, and Reporting Review the Considerations for Changing Field Types Review the Considerations for Modifying Data Models with Schema Builder Review the Considerations for Importing and Exporting Data Identify Use Cases of External Objects Determine an Appropriate Data Model Understand Design Implications with Complex Environments and Large Data Volumes (LDV) Data Management Review the Considerations for Working with LDV Review Data Lifecycle Concepts and Mechanisms Review Master Data Management and System of Record Concepts Review Data Migration, Planning, Preparation, and Execution Identify Potential LDV and Calculate Expected Volumes Go Further with Indexes Standard and Custom Skinny Tables Lock Records Security Model Review the Considerations for Working with Internal and External Sharing Models Restrict and Extend Object and Field Access Determine Sharing Solutions Identify Record Sharing Mechanisms (Declarative, Programmatic, Implicit) Understand Teams Concepts (Account, Opportunity, Case) Understand Person Accounts and Its Implication on Sharing Encrypt Data Sharing in Communities Understand Community Security Mechanisms Secure Integration Endpoints Integrate and Specify Connected Apps and Named Credentials Advanced Security & Visibility Concepts Implement Security & Visibility Controls with Apex and Visualforce Review Territory Management and Its Implication on Data Management, Sharing, and Visibility Review Divisions and Its Implication on Data Management, Sharing, and Visibility Understand Security and Visibility Controls on 'Special' Objects Solution Design Determine When to Leverage Standard Products Functionality vs. Custom Build vs. AppExchange Understand Declarative and Programmatic Configuration Get to Know the Order of Execution Automate Business Processes Consider Reporting and Analytics Needs Consider How to Store and Access Content/Files Apply Solution Design Concepts to Real-World Problems and Scenarios Deployment & Integration Best Practices Review the Application Lifecycle Understand How Sandboxes Should Be Used Review Deployment Options Identify Integration Patterns Wrap-up Review a Practice Scenario Review What Was Covered Additional course details: Nexus Humans Salesforce Build Application Architect Expertise (ARC901) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Salesforce Build Application Architect Expertise (ARC901) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Solutions architects, security DevOps, and security engineers Overview In this course, you will learn to: Establish a landing zone with AWS Control Tower Configure AWS Organizations to create a multi-account environment Implement identity management using AWS Single Sign-On users and groups Federate access using AWS SSO Enforce policies using prepackaged guardrails Centralize logging using AWS CloudTrail and AWS Config Enable cross-account security audits using AWS Identity and Access Management (IAM) Define workflows for provisioning accounts using AWS Service Catalog and AWS Security Hub Security is foundational to AWS. Governance at scale is a new concept for automating cloud governance that can help companies retire manual processes in account management, budget enforcement, and security and compliance. By automating common challenges, companies can scale without inhibiting agility, speed, or innovation. In addition, they can provide decision makers with the visibility, control, and governance necessary to protect sensitive data and systems.In this course, you will learn how to facilitate developer speed and agility, and incorporate preventive and detective controls. By the end of this course, you will be able to apply governance best practices. Course Introduction Instructor introduction Learning objectives Course structure and objectives Course logistics and agenda Module 1: Governance at Scale Governance at scale focal points Business and Technical Challenges Module 2: Governance Automation Multi-account strategies, guidance, and architecture Environments for agility and governance at scale Governance with AWS Control Tower Use cases for governance at scale Module 3: Preventive Controls Enterprise environment challenges for developers AWS Service Catalog Resource creation Workflows for provisioning accounts Preventive cost and security governance Self-service with existing IT service management (ITSM) tools Module 4: Detective Controls Operations aspect of governance at scale Resource monitoring Configuration rules for auditing Operational insights Remediation Clean up accounts Module 5: Resources Explore additional resources for security governance at scale Additional course details: Nexus Humans AWS Security Governance at Scale training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the AWS Security Governance at Scale course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 5 Days 30 CPD hours This course is intended for This intermediate course is for future DB2 for z/OS database administrators who need to acquire the basic skills required to administer a DB2 database. Overview Implement a DB2 database designUse database utilities to load and reorganize dataDefine and implement a DB2 database recovery strategyControl access to database using DB2 authorization facilities This course provides students with instruction on how to physically implement a logical database design in DB2. The course includes instruction on DB2 data management, DB2 catalog tables, the bind process, database utilities, & security considerations. Course Outline Setting up a DB2 database Referential integrity Getting data into and out of DB2 Keeping your DB2 data in good shape Application data recovery basics Program preparation and Bind Security Serialization Additional course details: Nexus Humans CV832 IBM DB2 11 for z/OS Database Administration Workshop Part 1 training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the CV832 IBM DB2 11 for z/OS Database Administration Workshop Part 1 course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 2 Days 12 CPD hours This course is intended for If you are a data analyst, data scientist, or a business analyst who wants to get started with using Python and machine learning techniques to analyze data and predict outcomes, this book is for you. Basic knowledge of computer programming and data analytics is a must. Familiarity with mathematical concepts such as algebra and basic statistics will be useful. Overview By the end of this course, you will have the skills you need to confidently use various machine learning algorithms to perform detailed data analysis and extract meaningful insights from data. This course is designed to give you practical guidance on industry-standard data analysis and machine learning tools in Python, with the help of realistic data. The course will help you understand how you can use pandas and Matplotlib to critically examine a dataset with summary statistics and graphs, and extract the insights you seek to derive. You will continue to build on your knowledge as you learn how to prepare data and feed it to machine learning algorithms, such as regularized logistic regression and random forest, using the scikit-learn package. You?ll discover how to tune the algorithms to provide the best predictions on new and unseen data. As you delve into later sections, you?ll be able to understand the working and output of these algorithms and gain insight into not only the predictive capabilities of the models but also their reasons for making these predictions. Data Exploration and Cleaning Python and the Anaconda Package Management System Different Types of Data Science Problems Loading the Case Study Data with Jupyter and pandas Data Quality Assurance and Exploration Exploring the Financial History Features in the Dataset Activity 1: Exploring Remaining Financial Features in the Dataset Introduction to Scikit-Learn and Model Evaluation Introduction Model Performance Metrics for Binary Classification Activity 2: Performing Logistic Regression with a New Feature and Creating a Precision-Recall Curve Details of Logistic Regression and Feature Exploration Introduction Examining the Relationships between Features and the Response Univariate Feature Selection: What It Does and Doesn't Do Building Cloud-Native Applications Activity 3: Fitting a Logistic Regression Model and Directly Using the Coefficients The Bias-Variance Trade-off Introduction Estimating the Coefficients and Intercepts of Logistic Regression Cross Validation: Choosing the Regularization Parameter and Other Hyperparameters Activity 4: Cross-Validation and Feature Engineering with the Case Study Data Decision Trees and Random Forests Introduction Decision trees Random Forests: Ensembles of Decision Trees Activity 5: Cross-Validation Grid Search with Random Forest Imputation of Missing Data, Financial Analysis, and Delivery to Client Introduction Review of Modeling Results Dealing with Missing Data: Imputation Strategies Activity 6: Deriving Financial Insights Final Thoughts on Delivering the Predictive Model to the Client
Duration 4 Days 24 CPD hours This course is intended for The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview Overview of data science and machine learning at scale Overview of the Hadoop ecosystem Working with HDFS data and Hive tables using Hue Introduction to Cloudera Data Science Workbench Overview of Apache Spark 2 Reading and writing data Inspecting data quality Cleansing and transforming data Summarizing and grouping data Combining, splitting, and reshaping data Exploring data Configuring, monitoring, and troubleshooting Spark applications Overview of machine learning in Spark MLlib Extracting, transforming, and selecting features Building and evaluating regression models Building and evaluating classification models Building and evaluating clustering models Cross-validating models and tuning hyperparameters Building machine learning pipelines Deploying machine learning models Spark, Spark SQL, and Spark MLlib PySpark and sparklyr Cloudera Data Science Workbench (CDSW) Hue This workshop covers data science and machine learning workflows at scale using Apache Spark 2 and other key components of the Hadoop ecosystem. The workshop emphasizes the use of data science and machine learning methods to address real-world business challenges. Using scenarios and datasets from a fictional technology company, students discover insights to support critical business decisions and develop data products to transform the business. The material is presented through a sequence of brief lectures, interactive demonstrations, extensive hands-on exercises, and discussions. The Apache Spark demonstrations and exercises are conducted in Python (with PySpark) and R (with sparklyr) using the Cloudera Data Science Workbench (CDSW) environment. The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview of data science and machine learning at scaleOverview of the Hadoop ecosystemWorking with HDFS data and Hive tables using HueIntroduction to Cloudera Data Science WorkbenchOverview of Apache Spark 2Reading and writing dataInspecting data qualityCleansing and transforming dataSummarizing and grouping dataCombining, splitting, and reshaping dataExploring dataConfiguring, monitoring, and troubleshooting Spark applicationsOverview of machine learning in Spark MLlibExtracting, transforming, and selecting featuresBuilding and evauating regression modelsBuilding and evaluating classification modelsBuilding and evaluating clustering modelsCross-validating models and tuning hyperparametersBuilding machine learning pipelinesDeploying machine learning models Additional course details: Nexus Humans Cloudera Data Scientist Training training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Cloudera Data Scientist Training course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 3 Days 18 CPD hours This course is intended for Data Wrangling with Python takes a practical approach to equip beginners with the most essential data analysis tools in the shortest possible time. It contains multiple activities that use real-life business scenarios for you to practice and apply your new skills in a highly relevant context. Overview By the end of this course, you will be confident in using a diverse array of sources to extract, clean, transform, and format your data efficiently. In this course you will start with the absolute basics of Python, focusing mainly on data structures. Then you will delve into the fundamental tools of data wrangling like NumPy and Pandas libraries. You'll explore useful insights into why you should stay away from traditional ways of data cleaning, as done in other languages, and take advantage of the specialized pre-built routines in Python.This combination of Python tips and tricks will also demonstrate how to use the same Python backend and extract/transform data from an array of sources including the Internet, large database vaults, and Excel financial tables. To help you prepare for more challenging scenarios, you'll cover how to handle missing or wrong data, and reformat it based on the requirements from the downstream analytics tool. The course will further help you grasp concepts through real-world examples and datasets. Introduction to Data Structure using Python Python for Data Wrangling Lists, Sets, Strings, Tuples, and Dictionaries Advanced Operations on Built-In Data Structure Advanced Data Structures Basic File Operations in Python Introduction to NumPy, Pandas, and Matplotlib NumPy Arrays Pandas DataFrames Statistics and Visualization with NumPy and Pandas Using NumPy and Pandas to Calculate Basic Descriptive Statistics on the DataFrame Deep Dive into Data Wrangling with Python Subsetting, Filtering, and Grouping Detecting Outliers and Handling Missing Values Concatenating, Merging, and Joining Useful Methods of Pandas Get Comfortable with a Different Kind of Data Sources Reading Data from Different Text-Based (and Non-Text-Based) Sources Introduction to BeautifulSoup4 and Web Page Parsing Learning the Hidden Secrets of Data Wrangling Advanced List Comprehension and the zip Function Data Formatting Advanced Web Scraping and Data Gathering Basics of Web Scraping and BeautifulSoup libraries Reading Data from XML RDBMS and SQL Refresher of RDBMS and SQL Using an RDBMS (MySQL/PostgreSQL/SQLite) Application in real life and Conclusion of course Applying Your Knowledge to a Real-life Data Wrangling Task An Extension to Data Wrangling
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: A technical audience at an intermediate level Overview Using Amazon SageMaker, this course teaches you how to: Prepare a dataset for training. Train and evaluate a machine learning model. Automatically tune a machine learning model. Prepare a machine learning model for production. Think critically about machine learning model results In this course, learn how to solve a real-world use case with machine learning and produce actionable results using Amazon SageMaker. This course teaches you how to use Amazon SageMaker to cover the different stages of the typical data science process, from analyzing and visualizing a data set, to preparing the data and feature engineering, down to the practical aspects of model building, training, tuning and deployment. Day 1 Business problem: Churn prediction Load and display the dataset Assess features and determine which Amazon SageMaker algorithm to use Use Amazon Sagemaker to train, evaluate, and automatically tune the model Deploy the model Assess relative cost of errors Additional course details: Nexus Humans Practical Data Science with Amazon SageMaker training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Practical Data Science with Amazon SageMaker course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.