• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

145 Notebook courses in Cardiff delivered Online

Python for Data Science: Hands-on Technical Overview (TTPS4873)

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for This introductory-level course is intended for Business Analysts and Data Analysts (or anyone else in the data science realm) who are already comfortable working with numerical data in Excel or other spreadsheet environments. No prior programming experience is required, and a browser is the only tool necessary for the course. Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Throughout the hands-on course students, will learn to leverage Python scripting for data science (to a basic level) using the most current and efficient skills and techniques. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore (to a basic level): How to work with Python interactively in web notebooks The essentials of Python scripting Key concepts necessary to enter the world of Data Science via Python This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it?s often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. An Overview of Python Why Python? Python in the Shell Python in Web Notebooks (iPython, Jupyter, Zeppelin) Demo: Python, Notebooks, and Data Science Getting Started Using variables Builtin functions Strings Numbers Converting among types Writing to the screen Command line parameters Flow Control About flow control White space Conditional expressions Relational and Boolean operators While loops Alternate loop exits Sequences, Arrays, Dictionaries and Sets About sequences Lists and list methods Tuples Indexing and slicing Iterating through a sequence Sequence functions, keywords, and operators List comprehensions Generator Expressions Nested sequences Working with Dictionaries Working with Sets Working with files File overview Opening a text file Reading a text file Writing to a text file Reading and writing raw (binary) data Functions Defining functions Parameters Global and local scope Nested functions Returning values Essential Demos Sorting Exceptions Importing Modules Classes Regular Expressions The standard library Math functions The string module Dates and times Working with dates and times Translating timestamps Parsing dates from text Formatting dates Calendar data Python and Data Science Data Science Essentials Pandas Overview NumPy Overview SciKit Overview MatPlotLib Overview Working with Python in Data Science Additional course details: Nexus Humans Python for Data Science: Hands-on Technical Overview (TTPS4873) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Python for Data Science: Hands-on Technical Overview (TTPS4873) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Python for Data Science: Hands-on Technical Overview (TTPS4873)
Delivered OnlineFlexible Dates
Price on Enquiry

Python for Data Science Primer: Hands-on Technical Overview (TTPS4872)

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for This introductory-level course is intended for Business Analysts and Data Analysts (or anyone else in the data science realm) who are already comfortable working with numerical data in Excel or other spreadsheet environments. No prior programming experience is required, and a browser is the only tool necessary for the course. Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Throughout the hands-on course students, will learn to leverage Python scripting for data science (to a basic level) using the most current and efficient skills and techniques. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore (to a basic level): How to work with Python interactively in web notebooks The essentials of Python scripting Key concepts necessary to enter the world of Data Science via Python This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it?s often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it's often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. Additional course details: Nexus Humans Python for Data Science Primer: Hands-on Technical Overview (TTPS4872) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Python for Data Science Primer: Hands-on Technical Overview (TTPS4872) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Python for Data Science Primer: Hands-on Technical Overview (TTPS4872)
Delivered OnlineFlexible Dates
Price on Enquiry

Building Batch Data Analytics Solutions on AWS

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Architects and operators who build and manage data analytics pipelines Overview In this course, you will learn to: Compare the features and benefits of data warehouses, data lakes, and modern data architectures Design and implement a batch data analytics solution Identify and apply appropriate techniques, including compression, to optimize data storage Select and deploy appropriate options to ingest, transform, and store data Choose the appropriate instance and node types, clusters, auto scaling, and network topology for a particular business use case Understand how data storage and processing affect the analysis and visualization mechanisms needed to gain actionable business insights Secure data at rest and in transit Monitor analytics workloads to identify and remediate problems Apply cost management best practices In this course, you will learn to build batch data analytics solutions using Amazon EMR, an enterprise-grade Apache Spark and Apache Hadoop managed service. You will learn how Amazon EMR integrates with open-source projects such as Apache Hive, Hue, and HBase, and with AWS services such as AWS Glue and AWS Lake Formation. The course addresses data collection, ingestion, cataloging, storage, and processing components in the context of Spark and Hadoop. You will learn to use EMR Notebooks to support both analytics and machine learning workloads. You will also learn to apply security, performance, and cost management best practices to the operation of Amazon EMR. Module A: Overview of Data Analytics and the Data Pipeline Data analytics use cases Using the data pipeline for analytics Module 1: Introduction to Amazon EMR Using Amazon EMR in analytics solutions Amazon EMR cluster architecture Interactive Demo 1: Launching an Amazon EMR cluster Cost management strategies Module 2: Data Analytics Pipeline Using Amazon EMR: Ingestion and Storage Storage optimization with Amazon EMR Data ingestion techniques Module 3: High-Performance Batch Data Analytics Using Apache Spark on Amazon EMR Apache Spark on Amazon EMR use cases Why Apache Spark on Amazon EMR Spark concepts Interactive Demo 2: Connect to an EMR cluster and perform Scala commands using the Spark shell Transformation, processing, and analytics Using notebooks with Amazon EMR Practice Lab 1: Low-latency data analytics using Apache Spark on Amazon EMR Module 4: Processing and Analyzing Batch Data with Amazon EMR and Apache Hive Using Amazon EMR with Hive to process batch data Transformation, processing, and analytics Practice Lab 2: Batch data processing using Amazon EMR with Hive Introduction to Apache HBase on Amazon EMR Module 5: Serverless Data Processing Serverless data processing, transformation, and analytics Using AWS Glue with Amazon EMR workloads Practice Lab 3: Orchestrate data processing in Spark using AWS Step Functions Module 6: Security and Monitoring of Amazon EMR Clusters Securing EMR clusters Interactive Demo 3: Client-side encryption with EMRFS Monitoring and troubleshooting Amazon EMR clusters Demo: Reviewing Apache Spark cluster history Module 7: Designing Batch Data Analytics Solutions Batch data analytics use cases Activity: Designing a batch data analytics workflow Module B: Developing Modern Data Architectures on AWS Modern data architectures

Building Batch Data Analytics Solutions on AWS
Delivered OnlineFlexible Dates
Price on Enquiry

The Machine Learning Pipeline on AWS

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is intended for: Developers Solutions Architects Data Engineers Anyone with little to no experience with ML and wants to learn about the ML pipeline using Amazon SageMaker Overview In this course, you will learn to: Select and justify the appropriate ML approach for a given business problem Use the ML pipeline to solve a specific business problem Train, evaluate, deploy, and tune an ML model using Amazon SageMaker Describe some of the best practices for designing scalable, cost-optimized, and secure ML pipelines in AWS Apply machine learning to a real-life business problem after the course is complete This course explores how to use the machine learning (ML) pipeline to solve a real business problem in a project-based learning environment. Students will learn about each phase of the pipeline from instructor presentations and demonstrations and then apply that knowledge to complete a project solving one of three business problems: fraud detection, recommendation engines, or flight delays. By the end of the course, students will have successfully built, trained, evaluated, tuned, and deployed an ML model using Amazon SageMaker that solves their selected business problem. Module 0: Introduction Pre-assessment Module 1: Introduction to Machine Learning and the ML Pipeline Overview of machine learning, including use cases, types of machine learning, and key concepts Overview of the ML pipeline Introduction to course projects and approach Module 2: Introduction to Amazon SageMaker Introduction to Amazon SageMaker Demo: Amazon SageMaker and Jupyter notebooks Hands-on: Amazon SageMaker and Jupyter notebooks Module 3: Problem Formulation Overview of problem formulation and deciding if ML is the right solution Converting a business problem into an ML problem Demo: Amazon SageMaker Ground Truth Hands-on: Amazon SageMaker Ground Truth Practice problem formulation Formulate problems for projects Module 4: Preprocessing Overview of data collection and integration, and techniques for data preprocessing and visualization Practice preprocessing Preprocess project data Class discussion about projects Module 5: Model Training Choosing the right algorithm Formatting and splitting your data for training Loss functions and gradient descent for improving your model Demo: Create a training job in Amazon SageMaker Module 6: Model Evaluation How to evaluate classification models How to evaluate regression models Practice model training and evaluation Train and evaluate project models Initial project presentations Module 7: Feature Engineering and Model Tuning Feature extraction, selection, creation, and transformation Hyperparameter tuning Demo: SageMaker hyperparameter optimization Practice feature engineering and model tuning Apply feature engineering and model tuning to projects Final project presentations Module 8: Deployment How to deploy, inference, and monitor your model on Amazon SageMaker Deploying ML at the edge Demo: Creating an Amazon SageMaker endpoint Post-assessment Course wrap-up Additional course details: Nexus Humans The Machine Learning Pipeline on AWS training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the The Machine Learning Pipeline on AWS course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

The Machine Learning Pipeline on AWS
Delivered OnlineFlexible Dates
Price on Enquiry

From Data to Insights with Google Cloud Platform

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for Data Analysts, Business Analysts, Business Intelligence professionals Cloud Data Engineers who will be partnering with Data Analysts to build scalable data solutions on Google Cloud Platform Overview This course teaches students the following skills: Derive insights from data using the analysis and visualization tools on Google Cloud Platform Interactively query datasets using Google BigQuery Load, clean, and transform data at scale Visualize data using Google Data Studio and other third-party platforms Distinguish between exploratory and explanatory analytics and when to use each approach Explore new datasets and uncover hidden insights quickly and effectively Optimizing data models and queries for price and performance Want to know how to query and process petabytes of data in seconds? Curious about data analysis that scales automatically as your data grows? Welcome to the Data Insights course! This four-course accelerated online specialization teaches course participants how to derive insights through data analysis and visualization using the Google Cloud Platform. The courses feature interactive scenarios and hands-on labs where participants explore, mine, load, visualize, and extract insights from diverse Google BigQuery datasets. The courses also cover data loading, querying, schema modeling, optimizing performance, query pricing, and data visualization. This specialization is intended for the following participants: Data Analysts, Business Analysts, Business Intelligence professionals Cloud Data Engineers who will be partnering with Data Analysts to build scalable data solutions on Google Cloud Platform To get the most out of this specialization, we recommend participants have some proficiency with ANSI SQL. Introduction to Data on the Google Cloud Platform Highlight Analytics Challenges Faced by Data Analysts Compare Big Data On-Premises vs on the Cloud Learn from Real-World Use Cases of Companies Transformed through Analytics on the Cloud Navigate Google Cloud Platform Project Basics Lab: Getting started with Google Cloud Platform Big Data Tools Overview Walkthrough Data Analyst Tasks, Challenges, and Introduce Google Cloud Platform Data Tools Demo: Analyze 10 Billion Records with Google BigQuery Explore 9 Fundamental Google BigQuery Features Compare GCP Tools for Analysts, Data Scientists, and Data Engineers Lab: Exploring Datasets with Google BigQuery Exploring your Data with SQL Compare Common Data Exploration Techniques Learn How to Code High Quality Standard SQL Explore Google BigQuery Public Datasets Visualization Preview: Google Data Studio Lab: Troubleshoot Common SQL Errors Google BigQuery Pricing Walkthrough of a BigQuery Job Calculate BigQuery Pricing: Storage, Querying, and Streaming Costs Optimize Queries for Cost Lab: Calculate Google BigQuery Pricing Cleaning and Transforming your Data Examine the 5 Principles of Dataset Integrity Characterize Dataset Shape and Skew Clean and Transform Data using SQL Clean and Transform Data using a new UI: Introducing Cloud Dataprep Lab: Explore and Shape Data with Cloud Dataprep Storing and Exporting Data Compare Permanent vs Temporary Tables Save and Export Query Results Performance Preview: Query Cache Lab: Creating new Permanent Tables Ingesting New Datasets into Google BigQuery Query from External Data Sources Avoid Data Ingesting Pitfalls Ingest New Data into Permanent Tables Discuss Streaming Inserts Lab: Ingesting and Querying New Datasets Data Visualization Overview of Data Visualization Principles Exploratory vs Explanatory Analysis Approaches Demo: Google Data Studio UI Connect Google Data Studio to Google BigQuery Lab: Exploring a Dataset in Google Data Studio Joining and Merging Datasets Merge Historical Data Tables with UNION Introduce Table Wildcards for Easy Merges Review Data Schemas: Linking Data Across Multiple Tables Walkthrough JOIN Examples and Pitfalls Lab: Join and Union Data from Multiple Tables Advanced Functions and Clauses Review SQL Case Statements Introduce Analytical Window Functions Safeguard Data with One-Way Field Encryption Discuss Effective Sub-query and CTE design Compare SQL and Javascript UDFs Lab: Deriving Insights with Advanced SQL Functions Schema Design and Nested Data Structures Compare Google BigQuery vs Traditional RDBMS Data Architecture Normalization vs Denormalization: Performance Tradeoffs Schema Review: The Good, The Bad, and The Ugly Arrays and Nested Data in Google BigQuery Lab: Querying Nested and Repeated Data More Visualization with Google Data Studio Create Case Statements and Calculated Fields Avoid Performance Pitfalls with Cache considerations Share Dashboards and Discuss Data Access considerations Optimizing for Performance Avoid Google BigQuery Performance Pitfalls Prevent Hotspots in your Data Diagnose Performance Issues with the Query Explanation map Lab: Optimizing and Troubleshooting Query Performance Advanced Insights Introducing Cloud Datalab Cloud Datalab Notebooks and Cells Benefits of Cloud Datalab Data Access Compare IAM and BigQuery Dataset Roles Avoid Access Pitfalls Review Members, Roles, Organizations, Account Administration, and Service Accounts

From Data to Insights with Google Cloud Platform
Delivered OnlineFlexible Dates
Price on Enquiry