Duration 2 Days 12 CPD hours This course is intended for If you are a data analyst, data scientist, or a business analyst who wants to get started with using Python and machine learning techniques to analyze data and predict outcomes, this book is for you. Basic knowledge of computer programming and data analytics is a must. Familiarity with mathematical concepts such as algebra and basic statistics will be useful. Overview By the end of this course, you will have the skills you need to confidently use various machine learning algorithms to perform detailed data analysis and extract meaningful insights from data. This course is designed to give you practical guidance on industry-standard data analysis and machine learning tools in Python, with the help of realistic data. The course will help you understand how you can use pandas and Matplotlib to critically examine a dataset with summary statistics and graphs, and extract the insights you seek to derive. You will continue to build on your knowledge as you learn how to prepare data and feed it to machine learning algorithms, such as regularized logistic regression and random forest, using the scikit-learn package. You?ll discover how to tune the algorithms to provide the best predictions on new and unseen data. As you delve into later sections, you?ll be able to understand the working and output of these algorithms and gain insight into not only the predictive capabilities of the models but also their reasons for making these predictions. Data Exploration and Cleaning Python and the Anaconda Package Management System Different Types of Data Science Problems Loading the Case Study Data with Jupyter and pandas Data Quality Assurance and Exploration Exploring the Financial History Features in the Dataset Activity 1: Exploring Remaining Financial Features in the Dataset Introduction to Scikit-Learn and Model Evaluation Introduction Model Performance Metrics for Binary Classification Activity 2: Performing Logistic Regression with a New Feature and Creating a Precision-Recall Curve Details of Logistic Regression and Feature Exploration Introduction Examining the Relationships between Features and the Response Univariate Feature Selection: What It Does and Doesn't Do Building Cloud-Native Applications Activity 3: Fitting a Logistic Regression Model and Directly Using the Coefficients The Bias-Variance Trade-off Introduction Estimating the Coefficients and Intercepts of Logistic Regression Cross Validation: Choosing the Regularization Parameter and Other Hyperparameters Activity 4: Cross-Validation and Feature Engineering with the Case Study Data Decision Trees and Random Forests Introduction Decision trees Random Forests: Ensembles of Decision Trees Activity 5: Cross-Validation Grid Search with Random Forest Imputation of Missing Data, Financial Analysis, and Delivery to Client Introduction Review of Modeling Results Dealing with Missing Data: Imputation Strategies Activity 6: Deriving Financial Insights Final Thoughts on Delivering the Predictive Model to the Client
Duration 4 Days 24 CPD hours This course is intended for The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview Overview of data science and machine learning at scale Overview of the Hadoop ecosystem Working with HDFS data and Hive tables using Hue Introduction to Cloudera Data Science Workbench Overview of Apache Spark 2 Reading and writing data Inspecting data quality Cleansing and transforming data Summarizing and grouping data Combining, splitting, and reshaping data Exploring data Configuring, monitoring, and troubleshooting Spark applications Overview of machine learning in Spark MLlib Extracting, transforming, and selecting features Building and evaluating regression models Building and evaluating classification models Building and evaluating clustering models Cross-validating models and tuning hyperparameters Building machine learning pipelines Deploying machine learning models Spark, Spark SQL, and Spark MLlib PySpark and sparklyr Cloudera Data Science Workbench (CDSW) Hue This workshop covers data science and machine learning workflows at scale using Apache Spark 2 and other key components of the Hadoop ecosystem. The workshop emphasizes the use of data science and machine learning methods to address real-world business challenges. Using scenarios and datasets from a fictional technology company, students discover insights to support critical business decisions and develop data products to transform the business. The material is presented through a sequence of brief lectures, interactive demonstrations, extensive hands-on exercises, and discussions. The Apache Spark demonstrations and exercises are conducted in Python (with PySpark) and R (with sparklyr) using the Cloudera Data Science Workbench (CDSW) environment. The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview of data science and machine learning at scaleOverview of the Hadoop ecosystemWorking with HDFS data and Hive tables using HueIntroduction to Cloudera Data Science WorkbenchOverview of Apache Spark 2Reading and writing dataInspecting data qualityCleansing and transforming dataSummarizing and grouping dataCombining, splitting, and reshaping dataExploring dataConfiguring, monitoring, and troubleshooting Spark applicationsOverview of machine learning in Spark MLlibExtracting, transforming, and selecting featuresBuilding and evauating regression modelsBuilding and evaluating classification modelsBuilding and evaluating clustering modelsCross-validating models and tuning hyperparametersBuilding machine learning pipelinesDeploying machine learning models Additional course details: Nexus Humans Cloudera Data Scientist Training training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Cloudera Data Scientist Training course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: A technical audience at an intermediate level Overview Using Amazon SageMaker, this course teaches you how to: Prepare a dataset for training. Train and evaluate a machine learning model. Automatically tune a machine learning model. Prepare a machine learning model for production. Think critically about machine learning model results In this course, learn how to solve a real-world use case with machine learning and produce actionable results using Amazon SageMaker. This course teaches you how to use Amazon SageMaker to cover the different stages of the typical data science process, from analyzing and visualizing a data set, to preparing the data and feature engineering, down to the practical aspects of model building, training, tuning and deployment. Day 1 Business problem: Churn prediction Load and display the dataset Assess features and determine which Amazon SageMaker algorithm to use Use Amazon Sagemaker to train, evaluate, and automatically tune the model Deploy the model Assess relative cost of errors Additional course details: Nexus Humans Practical Data Science with Amazon SageMaker training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Practical Data Science with Amazon SageMaker course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 2 Days 12 CPD hours This course is intended for This introductory-level course is intended for Business Analysts and Data Analysts (or anyone else in the data science realm) who are already comfortable working with numerical data in Excel or other spreadsheet environments. No prior programming experience is required, and a browser is the only tool necessary for the course. Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Our engaging instructors and mentors are highly experienced practitioners who bring years of current 'on-the-job' experience into every classroom. Throughout the hands-on course students, will learn to leverage Python scripting for data science (to a basic level) using the most current and efficient skills and techniques. Working in a hands-on learning environment, guided by our expert team, attendees will learn about and explore (to a basic level): How to work with Python interactively in web notebooks The essentials of Python scripting Key concepts necessary to enter the world of Data Science via Python This course introduces data analysts and business analysts (as well as anyone interested in Data Science) to the Python programming language, as it?s often used in Data Science in web notebooks. This goal of this course is to provide students with a baseline understanding of core concepts that can serve as a platform of knowledge to follow up with more in-depth training and real-world practice. An Overview of Python Why Python? Python in the Shell Python in Web Notebooks (iPython, Jupyter, Zeppelin) Demo: Python, Notebooks, and Data Science Getting Started Using variables Builtin functions Strings Numbers Converting among types Writing to the screen Command line parameters Flow Control About flow control White space Conditional expressions Relational and Boolean operators While loops Alternate loop exits Sequences, Arrays, Dictionaries and Sets About sequences Lists and list methods Tuples Indexing and slicing Iterating through a sequence Sequence functions, keywords, and operators List comprehensions Generator Expressions Nested sequences Working with Dictionaries Working with Sets Working with files File overview Opening a text file Reading a text file Writing to a text file Reading and writing raw (binary) data Functions Defining functions Parameters Global and local scope Nested functions Returning values Essential Demos Sorting Exceptions Importing Modules Classes Regular Expressions The standard library Math functions The string module Dates and times Working with dates and times Translating timestamps Parsing dates from text Formatting dates Calendar data Python and Data Science Data Science Essentials Pandas Overview NumPy Overview SciKit Overview MatPlotLib Overview Working with Python in Data Science Additional course details: Nexus Humans Python for Data Science: Hands-on Technical Overview (TTPS4873) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Python for Data Science: Hands-on Technical Overview (TTPS4873) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Architecture and course review Post course knowledge check Architecture review Course review
Duration 3 Days 18 CPD hours This course is intended for This course is intended for: Intermediate software developers Overview In this course, you will learn to: Set up the AWS SDK and developer credentials for Java, C#/.NET, Python, and JavaScript Interact with AWS services and develop solutions by using the AWS SDK Use AWS Identity and Access Management (IAM) for service authentication Use Amazon Simple Storage Service (Amazon S3) and Amazon DynamoDB as data stores Integrate applications and data by using AWS Lambda, Amazon API Gateway, Amazon Simple Queue Service (Amazon SQS), Amazon Simple Notification Service (Amazon SNS), and AWS Step Functions Use Amazon Cognito for user authentication Use Amazon ElastiCache to improve application scalability Leverage the CI/CD pipeline to deploy applications on AWS In this course, you learn how to use the AWS SDK to develop secure and scalable cloud applications using multiple AWS services such as Amazon DynamoDB, Amazon Simple Storage Service, and AWS Lambda. You explore how to interact with AWS using code and learn about key concepts, best practices, and troubleshooting tips. Module 0: Course Overview Agenda Introductions Student resources Module 1: Introduction to AWS Introduction to the AWS Cloud Cloud scenarios Infrastructure overview Introduction to AWS foundation services Module 2: Introduction to Developing on AWS Getting started with developing on AWS Introduction to developer tools Introduction to management tools Module 3: Introduction to AWS Identity and Access Management Shared responsibility model Introduction to IAM Use authentication and authorization Module 4: Introduction to the Lab Environment Introduction to the lab environment Lab 1: Getting started and working with IAM Module 5: Developing Storage Solutions with Amazon Simple Storage Service Overview of AWS storage options Amazon S3 key concepts Best practices Troubleshooting Scenario: Building a complete application Lab 2: Developing storage solutions with Amazon S3 Module 6: Developing Flexible NoSQL Solutions with Amazon DynamoDB Introduction to AWS database options Introduction to Amazon DynamoDB Developing with DynamoDB Best practices Troubleshooting Scenario: Building an end-to-end app Lab 3: Developing flexible NoSQL solutions with Amazon DynamoDB Module 7: Developing Event-Driven Solutions with AWS Lambda What is serverless computing? Introduction to AWS Lambda Key concepts How Lambda works Use cases Best practices Scenario: Build an end-to-end app Module 8: Developing Solutions with Amazon API Gateway Introduction to Amazon API Gateway Developing with API Gateway Best practices Introduction to AWS Serverless Application Model Scenario: Building an end-to-end app Lab 4: Developing event-driven solutions with AWS Lambda Module 9: Developing Solutions with AWS Step Functions Understanding the need for Step Functions Introduction to AWS Step Functions Use cases Module 10: Developing Solutions with Amazon Simple Queue Service and Amazon Simple Notification Service Why use a queueing service? Developing with Amazon Simple Queue Service Developing with Amazon Simple Notification Service Developing with Amazon MQ Lab 5: Developing messaging solutions with Amazon SQS and Amazon SNS Module 11: Caching Information with Amazon ElastiCache Caching overview Caching with Amazon ElastiCache Caching strategies Module 12: Developing Secure Applications Securing your applications Authenticating your applications to AWS Authenticating your customers Scenario: Building an end-to-end app Module 13: Deploying Applications Introduction to DevOps Introduction to deployment and testing strategies Deploying applications with AWS Elastic Beanstalk Scenario: Building an end-to-end app Lab 6: Building an end-to-end app Module 14: Course wrap-up Course overview AWS training courses Certifications Course feedback
Duration 1 Days 6 CPD hours In this hands on workshop for Agile Scrum Masters, Release Train Engineers and anyone serving as Jira Administrators, Jira experts will lead you through advanced configuration and customization settings in Jira, from installation through to customized screens, workflows, filters and reports. Jira Administration Adding and managing Users Administering and managing Groups Global Jira Settings Jira layout and interface customization User authentication and security Jira Customization Customization of screens and fields Customization of workflows Project and Board Administration Configuring and managing Projects Configuring and managing Boards Creating and managing Filters JQL Jira Integration Integrating Jira with Atlassian Tools Retrospectives and Documentation in Confluence Code management with Bitbucket Integration management with Bamboo Building a Dashboard with gadgets Jira Plug-ins and Marketplace
This course has a simple objective: to help gain appointments with potential clients. In most consultative selling situations clients won't commit to purchases over the telephone. This means setting up a meeting to discuss the options with them face-to-face. But getting 'face time' can be tricky. This practical workshop can help. Participants will acquire essential tools, skills and methods; discuss specific organisational issues; and identify areas for improvement. They will discover how to: Increase their effectiveness through proper preparation Construct attention-grabbing opening statements Help potential clients feel comfortable agreeing to a meeting Develop tactics for responding to difficult excuses and objections Stress the benefits of a face-to-face consultation Develop and enhance their questioning and listening skills Prevent customers cancelling booked appointments 1 Introduction to appointment setting Key trends that have changed the way people buy today - and will buy tomorrow Why many sales people avoid picking up the phone The difference that makes a difference - what makes a good appointment-maker? 2 Before you pick up the telephone It all starts with a plan... Who and what to focus our attention How much research should we undertake and why? Setting primary and secondary objectives 3 Making your approach Key considerations Every call is an opportunity - creating a positive mind-set Using a structured approach Using partnership language 4 Gaining an insight into the customer's needs How to quickly 'tune in' to your customers, so that you can serve them more easily Developing speech patterns that put customers at their ease Using effective questioning and listening skills Finding and building pain points 5 Dealing with excuses and objections Pre-empting potential excuses Developing techniques for responding to client objections Keeping the door open for future contact 6 Securing the appointment Selling the benefits of a consultancy meeting Techniques for avoiding cancelled appointments Gaining commitment 7 Action plans Course summary and presentation of action plans
Mindfulness is a practical technique for developing a greater sense of awareness and focus on the present moment. It is the opposite of mindlessness, meaning that actions and reactions become conscious and deliberate. It is an extremely useful tool for any busy work environment. Currently being used by the likes of Google and Pepsi, mindfulness can be adopted within the workplace to reduce stress and anxiety, provide greater focus and clarity, improve leadership capabilities and enhance the general wellbeing of staff at all levels. This workshop has been developed for forward-thinking organisations wanting to make a real and sustainable commitment to improving workplace wellbeing and productivity. This workshop will help you to understand the basic principles and benefits of mindfulness, and how it can be used in the workplace setting. It will also enable you to develop techniques to alleviate overwhelming feelings of stress or anxiety, prepare for important or challenging meetings, and generally achieve a greater sense of focus, clarity and calm whilst dealing with a hectic schedule.