Duration 1 Days 6 CPD hours This course is intended for Individuals taking this course are business professionals seeking to develop or increase their emotional intelligence. Overview In this course, you will explore the concept of emotional intelligence. You will: Identify the components of emotional intelligence and recognize how emotional intelligence benefits organizations. Assess and develop your personal emotional intelligence competencies. Assess and develop your social emotional intelligence competencies. Practice emotional intelligence in common workplace scenarios. It was once believed that intelligence was the metric that would determine a person's success in the workplace. Intelligence matters because it contributes to your ability to do your job. But intelligence is not the best indicator of whether or not you'll succeed. Your ability to understand and manage your own emotions, and get along well with others, has at least as much impact on your performance and effectiveness as intelligence. In this course, you'll explore strategies to increase your awareness of your emotions, develop your ability to manage your emotions, and improve your social skills. Recognizing the Benefits of Emotional Intelligence Define Emotional Intelligence Recognize EQ's Impact on Work Experience Increasing Your Personal Emotional Intelligence in the Workplace Develop Your Level of Self-Awareness Develop Your Self-Regulation Skills Develop Your Motivation Increasing Your Social Emotional Intelligence in the Workplace Develop Your Empathy Develop Your Social Skills Practicing Emotional Intelligence in the Workplace Practice Emotionally Intelligent Leadership Build an Emotionally Intelligent Team Manage Change Manage Conflict Coach for Performance Additional course details: Nexus Humans Emotional Intelligence for Business Professionals (Second Edition) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Emotional Intelligence for Business Professionals (Second Edition) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: A technical audience at an intermediate level Overview Using Amazon SageMaker, this course teaches you how to: Prepare a dataset for training. Train and evaluate a machine learning model. Automatically tune a machine learning model. Prepare a machine learning model for production. Think critically about machine learning model results In this course, learn how to solve a real-world use case with machine learning and produce actionable results using Amazon SageMaker. This course teaches you how to use Amazon SageMaker to cover the different stages of the typical data science process, from analyzing and visualizing a data set, to preparing the data and feature engineering, down to the practical aspects of model building, training, tuning and deployment. Day 1 Business problem: Churn prediction Load and display the dataset Assess features and determine which Amazon SageMaker algorithm to use Use Amazon Sagemaker to train, evaluate, and automatically tune the model Deploy the model Assess relative cost of errors Additional course details: Nexus Humans Practical Data Science with Amazon SageMaker training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Practical Data Science with Amazon SageMaker course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 3 Days 18 CPD hours This course is intended for Data Wrangling with Python takes a practical approach to equip beginners with the most essential data analysis tools in the shortest possible time. It contains multiple activities that use real-life business scenarios for you to practice and apply your new skills in a highly relevant context. Overview By the end of this course, you will be confident in using a diverse array of sources to extract, clean, transform, and format your data efficiently. In this course you will start with the absolute basics of Python, focusing mainly on data structures. Then you will delve into the fundamental tools of data wrangling like NumPy and Pandas libraries. You'll explore useful insights into why you should stay away from traditional ways of data cleaning, as done in other languages, and take advantage of the specialized pre-built routines in Python.This combination of Python tips and tricks will also demonstrate how to use the same Python backend and extract/transform data from an array of sources including the Internet, large database vaults, and Excel financial tables. To help you prepare for more challenging scenarios, you'll cover how to handle missing or wrong data, and reformat it based on the requirements from the downstream analytics tool. The course will further help you grasp concepts through real-world examples and datasets. Introduction to Data Structure using Python Python for Data Wrangling Lists, Sets, Strings, Tuples, and Dictionaries Advanced Operations on Built-In Data Structure Advanced Data Structures Basic File Operations in Python Introduction to NumPy, Pandas, and Matplotlib NumPy Arrays Pandas DataFrames Statistics and Visualization with NumPy and Pandas Using NumPy and Pandas to Calculate Basic Descriptive Statistics on the DataFrame Deep Dive into Data Wrangling with Python Subsetting, Filtering, and Grouping Detecting Outliers and Handling Missing Values Concatenating, Merging, and Joining Useful Methods of Pandas Get Comfortable with a Different Kind of Data Sources Reading Data from Different Text-Based (and Non-Text-Based) Sources Introduction to BeautifulSoup4 and Web Page Parsing Learning the Hidden Secrets of Data Wrangling Advanced List Comprehension and the zip Function Data Formatting Advanced Web Scraping and Data Gathering Basics of Web Scraping and BeautifulSoup libraries Reading Data from XML RDBMS and SQL Refresher of RDBMS and SQL Using an RDBMS (MySQL/PostgreSQL/SQLite) Application in real life and Conclusion of course Applying Your Knowledge to a Real-life Data Wrangling Task An Extension to Data Wrangling
Duration 2 Days 12 CPD hours This course is intended for If you are a data analyst, data scientist, or a business analyst who wants to get started with using Python and machine learning techniques to analyze data and predict outcomes, this book is for you. Basic knowledge of computer programming and data analytics is a must. Familiarity with mathematical concepts such as algebra and basic statistics will be useful. Overview By the end of this course, you will have the skills you need to confidently use various machine learning algorithms to perform detailed data analysis and extract meaningful insights from data. This course is designed to give you practical guidance on industry-standard data analysis and machine learning tools in Python, with the help of realistic data. The course will help you understand how you can use pandas and Matplotlib to critically examine a dataset with summary statistics and graphs, and extract the insights you seek to derive. You will continue to build on your knowledge as you learn how to prepare data and feed it to machine learning algorithms, such as regularized logistic regression and random forest, using the scikit-learn package. You?ll discover how to tune the algorithms to provide the best predictions on new and unseen data. As you delve into later sections, you?ll be able to understand the working and output of these algorithms and gain insight into not only the predictive capabilities of the models but also their reasons for making these predictions. Data Exploration and Cleaning Python and the Anaconda Package Management System Different Types of Data Science Problems Loading the Case Study Data with Jupyter and pandas Data Quality Assurance and Exploration Exploring the Financial History Features in the Dataset Activity 1: Exploring Remaining Financial Features in the Dataset Introduction to Scikit-Learn and Model Evaluation Introduction Model Performance Metrics for Binary Classification Activity 2: Performing Logistic Regression with a New Feature and Creating a Precision-Recall Curve Details of Logistic Regression and Feature Exploration Introduction Examining the Relationships between Features and the Response Univariate Feature Selection: What It Does and Doesn't Do Building Cloud-Native Applications Activity 3: Fitting a Logistic Regression Model and Directly Using the Coefficients The Bias-Variance Trade-off Introduction Estimating the Coefficients and Intercepts of Logistic Regression Cross Validation: Choosing the Regularization Parameter and Other Hyperparameters Activity 4: Cross-Validation and Feature Engineering with the Case Study Data Decision Trees and Random Forests Introduction Decision trees Random Forests: Ensembles of Decision Trees Activity 5: Cross-Validation Grid Search with Random Forest Imputation of Missing Data, Financial Analysis, and Delivery to Client Introduction Review of Modeling Results Dealing with Missing Data: Imputation Strategies Activity 6: Deriving Financial Insights Final Thoughts on Delivering the Predictive Model to the Client
Duration 4 Days 24 CPD hours This course is intended for The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview Overview of data science and machine learning at scale Overview of the Hadoop ecosystem Working with HDFS data and Hive tables using Hue Introduction to Cloudera Data Science Workbench Overview of Apache Spark 2 Reading and writing data Inspecting data quality Cleansing and transforming data Summarizing and grouping data Combining, splitting, and reshaping data Exploring data Configuring, monitoring, and troubleshooting Spark applications Overview of machine learning in Spark MLlib Extracting, transforming, and selecting features Building and evaluating regression models Building and evaluating classification models Building and evaluating clustering models Cross-validating models and tuning hyperparameters Building machine learning pipelines Deploying machine learning models Spark, Spark SQL, and Spark MLlib PySpark and sparklyr Cloudera Data Science Workbench (CDSW) Hue This workshop covers data science and machine learning workflows at scale using Apache Spark 2 and other key components of the Hadoop ecosystem. The workshop emphasizes the use of data science and machine learning methods to address real-world business challenges. Using scenarios and datasets from a fictional technology company, students discover insights to support critical business decisions and develop data products to transform the business. The material is presented through a sequence of brief lectures, interactive demonstrations, extensive hands-on exercises, and discussions. The Apache Spark demonstrations and exercises are conducted in Python (with PySpark) and R (with sparklyr) using the Cloudera Data Science Workbench (CDSW) environment. The workshop is designed for data scientists who currently use Python or R to work with smaller datasets on a single machine and who need to scale up their analyses and machine learning models to large datasets on distributed clusters. Data engineers and developers with some knowledge of data science and machine learning may also find this workshop useful. Overview of data science and machine learning at scaleOverview of the Hadoop ecosystemWorking with HDFS data and Hive tables using HueIntroduction to Cloudera Data Science WorkbenchOverview of Apache Spark 2Reading and writing dataInspecting data qualityCleansing and transforming dataSummarizing and grouping dataCombining, splitting, and reshaping dataExploring dataConfiguring, monitoring, and troubleshooting Spark applicationsOverview of machine learning in Spark MLlibExtracting, transforming, and selecting featuresBuilding and evauating regression modelsBuilding and evaluating classification modelsBuilding and evaluating clustering modelsCross-validating models and tuning hyperparametersBuilding machine learning pipelinesDeploying machine learning models Additional course details: Nexus Humans Cloudera Data Scientist Training training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Cloudera Data Scientist Training course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 5 Days 30 CPD hours This course is intended for This course is designed for business professionals who leverage data to address business issues. The typical student in this course will have several years of experience with computing technology, including some aptitude in computer programming. However, there is not necessarily a single organizational role that this course targets. A prospective student might be a programmer looking to expand their knowledge of how to guide business decisions by collecting, wrangling, analyzing, and manipulating data through code; or a data analyst with a background in applied math and statistics who wants to take their skills to the next level; or any number of other data-driven situations. Ultimately, the target student is someone who wants to learn how to more effectively extract insights from their work and leverage that insight in addressing business issues, thereby bringing greater value to the business. Overview In this course, you will learn to: Use data science principles to address business issues. Apply the extract, transform, and load (ETL) process to prepare datasets. Use multiple techniques to analyze data and extract valuable insights. Design a machine learning approach to address business issues. Train, tune, and evaluate classification models. Train, tune, and evaluate regression and forecasting models. Train, tune, and evaluate clustering models. Finalize a data science project by presenting models to an audience, putting models into production, and monitoring model performance. For a business to thrive in our data-driven world, it must treat data as one of its most important assets. Data is crucial for understanding where the business is and where it's headed. Not only can data reveal insights, it can also inform?by guiding decisions and influencing day-to-day operations. This calls for a robust workforce of professionals who can analyze, understand, manipulate, and present data within an effective and repeatable process framework. In other words, the business world needs data science practitioners. This course will enable you to bring value to the business by putting data science concepts into practice Addressing Business Issues with Data Science Topic A: Initiate a Data Science Project Topic B: Formulate a Data Science Problem Extracting, Transforming, and Loading Data Topic A: Extract Data Topic B: Transform Data Topic C: Load Data Analyzing Data Topic A: Examine Data Topic B: Explore the Underlying Distribution of Data Topic C: Use Visualizations to Analyze Data Topic D: Preprocess Data Designing a Machine Learning Approach Topic A: Identify Machine Learning Concepts Topic B: Test a Hypothesis Developing Classification Models Topic A: Train and Tune Classification Models Topic B: Evaluate Classification Models Developing Regression Models Topic A: Train and Tune Regression Models Topic B: Evaluate Regression Models Developing Clustering Models Topic A: Train and Tune Clustering Models Topic B: Evaluate Clustering Models Finalizing a Data Science Project Topic A: Communicate Results to Stakeholders Topic B: Demonstrate Models in a Web App Topic C: Implement and Test Production Pipelines
Duration 3 Days 18 CPD hours This course is intended for This course is intended for: Intermediate software developers Overview In this course, you will learn to: Set up the AWS SDK and developer credentials for Java, C#/.NET, Python, and JavaScript Interact with AWS services and develop solutions by using the AWS SDK Use AWS Identity and Access Management (IAM) for service authentication Use Amazon Simple Storage Service (Amazon S3) and Amazon DynamoDB as data stores Integrate applications and data by using AWS Lambda, Amazon API Gateway, Amazon Simple Queue Service (Amazon SQS), Amazon Simple Notification Service (Amazon SNS), and AWS Step Functions Use Amazon Cognito for user authentication Use Amazon ElastiCache to improve application scalability Leverage the CI/CD pipeline to deploy applications on AWS In this course, you learn how to use the AWS SDK to develop secure and scalable cloud applications using multiple AWS services such as Amazon DynamoDB, Amazon Simple Storage Service, and AWS Lambda. You explore how to interact with AWS using code and learn about key concepts, best practices, and troubleshooting tips. Module 0: Course Overview Agenda Introductions Student resources Module 1: Introduction to AWS Introduction to the AWS Cloud Cloud scenarios Infrastructure overview Introduction to AWS foundation services Module 2: Introduction to Developing on AWS Getting started with developing on AWS Introduction to developer tools Introduction to management tools Module 3: Introduction to AWS Identity and Access Management Shared responsibility model Introduction to IAM Use authentication and authorization Module 4: Introduction to the Lab Environment Introduction to the lab environment Lab 1: Getting started and working with IAM Module 5: Developing Storage Solutions with Amazon Simple Storage Service Overview of AWS storage options Amazon S3 key concepts Best practices Troubleshooting Scenario: Building a complete application Lab 2: Developing storage solutions with Amazon S3 Module 6: Developing Flexible NoSQL Solutions with Amazon DynamoDB Introduction to AWS database options Introduction to Amazon DynamoDB Developing with DynamoDB Best practices Troubleshooting Scenario: Building an end-to-end app Lab 3: Developing flexible NoSQL solutions with Amazon DynamoDB Module 7: Developing Event-Driven Solutions with AWS Lambda What is serverless computing? Introduction to AWS Lambda Key concepts How Lambda works Use cases Best practices Scenario: Build an end-to-end app Module 8: Developing Solutions with Amazon API Gateway Introduction to Amazon API Gateway Developing with API Gateway Best practices Introduction to AWS Serverless Application Model Scenario: Building an end-to-end app Lab 4: Developing event-driven solutions with AWS Lambda Module 9: Developing Solutions with AWS Step Functions Understanding the need for Step Functions Introduction to AWS Step Functions Use cases Module 10: Developing Solutions with Amazon Simple Queue Service and Amazon Simple Notification Service Why use a queueing service? Developing with Amazon Simple Queue Service Developing with Amazon Simple Notification Service Developing with Amazon MQ Lab 5: Developing messaging solutions with Amazon SQS and Amazon SNS Module 11: Caching Information with Amazon ElastiCache Caching overview Caching with Amazon ElastiCache Caching strategies Module 12: Developing Secure Applications Securing your applications Authenticating your applications to AWS Authenticating your customers Scenario: Building an end-to-end app Module 13: Deploying Applications Introduction to DevOps Introduction to deployment and testing strategies Deploying applications with AWS Elastic Beanstalk Scenario: Building an end-to-end app Lab 6: Building an end-to-end app Module 14: Course wrap-up Course overview AWS training courses Certifications Course feedback
Duration 1 Days 6 CPD hours In this hands on workshop for Agile Scrum Masters, Release Train Engineers and anyone serving as Jira Administrators, Jira experts will lead you through advanced configuration and customization settings in Jira, from installation through to customized screens, workflows, filters and reports. Jira Administration Adding and managing Users Administering and managing Groups Global Jira Settings Jira layout and interface customization User authentication and security Jira Customization Customization of screens and fields Customization of workflows Project and Board Administration Configuring and managing Projects Configuring and managing Boards Creating and managing Filters JQL Jira Integration Integrating Jira with Atlassian Tools Retrospectives and Documentation in Confluence Code management with Bitbucket Integration management with Bamboo Building a Dashboard with gadgets Jira Plug-ins and Marketplace
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Architecture and course review Post course knowledge check Architecture review Course review
Pushy boardroom bullies... battling agendas... conflicts of interest. Here are the most common drivers of trouble in the boardroom, and tools for making things right.