• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

194 Data Skills courses in Liverpool delivered Online

R Programming for Data Science (v1.0)

By Nexus Human

Duration 5 Days 30 CPD hours This course is intended for This course is designed for students who want to learn the R programming language, particularly students who want to leverage R for data analysis and data science tasks in their organization. The course is also designed for students with an interest in applying statistics to real-world problems. A typical student in this course should have several years of experience with computing technology, along with a proficiency in at least one other programming language. Overview In this course, you will use R to perform common data science tasks.You will: Set up an R development environment and execute simple code. Perform operations on atomic data types in R, including characters, numbers, and logicals. Perform operations on data structures in R, including vectors, lists, and data frames. Write conditional statements and loops. Structure code for reuse with functions and packages. Manage data by loading and saving datasets, manipulating data frames, and more. Analyze data through exploratory analysis, statistical analysis, and more. Create and format data visualizations using base R and ggplot2. Create simple statistical models from data. In our data-driven world, organizations need the right tools to extract valuable insights from that data. The R programming language is one of the tools at the forefront of data science. Its robust set of packages and statistical functions makes it a powerful choice for analyzing data, manipulating data, performing statistical tests on data, and creating predictive models from data. Likewise, R is notable for its strong data visualization tools, enabling you to create high-quality graphs and plots that are incredibly customizable. This course will teach you the fundamentals of programming in R to get you started. It will also teach you how to use R to perform common data science tasks and achieve data-driven results for the business. Lesson 1: Setting Up R and Executing Simple Code Topic A: Set Up the R Development Environment Topic B: Write R Statements Lesson 2: Processing Atomic Data Types Topic A: Process Characters Topic B: Process Numbers Topic C: Process Logicals Lesson 3: Processing Data Structures Topic A: Process Vectors Topic B: Process Factors Topic C: Process Data Frames Topic D: Subset Data Structures Lesson 4: Writing Conditional Statements and Loops Topic A: Write Conditional Statements Topic B: Write Loops Lesson 5: Structuring Code for Reuse Topic A: Define and Call Functions Topic B: Apply Loop Functions Topic C: Manage R Packages Lesson 6: Managing Data in R Topic A: Load Data Topic B: Save Data Topic C: Manipulate Data Frames Using Base R Topic D: Manipulate Data Frames Using dplyr Topic E: Handle Dates and Times Lesson 7: Analyzing Data in R Topic A: Examine Data Topic B: Explore the Underlying Distribution of Data Topic C: Identify Missing Values Lesson 8: Visualizing Data in R Topic A: Plot Data Using Base R Functions Topic B: Plot Data Using ggplot2 Topic C: Format Plots in ggplot2 Topic D: Create Combination Plots Lesson 9: Modeling Data in R Topic A: Create Statistical Models in R Topic B: Create Machine Learning Models in R

R Programming for Data Science (v1.0)
Delivered OnlineFlexible Dates
Price on Enquiry

Python With Data Science

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for Audience: Data Scientists, Software Developers, IT Architects, and Technical Managers. Participants should have the general knowledge of statistics and programming Also familiar with Python Overview ? NumPy, pandas, Matplotlib, scikit-learn ? Python REPLs ? Jupyter Notebooks ? Data analytics life-cycle phases ? Data repairing and normalizing ? Data aggregation and grouping ? Data visualization ? Data science algorithms for supervised and unsupervised machine learning Covers theoretical and technical aspects of using Python in Applied Data Science projects and Data Logistics use cases. Python for Data Science ? Using Modules ? Listing Methods in a Module ? Creating Your Own Modules ? List Comprehension ? Dictionary Comprehension ? String Comprehension ? Python 2 vs Python 3 ? Sets (Python 3+) ? Python Idioms ? Python Data Science ?Ecosystem? ? NumPy ? NumPy Arrays ? NumPy Idioms ? pandas ? Data Wrangling with pandas' DataFrame ? SciPy ? Scikit-learn ? SciPy or scikit-learn? ? Matplotlib ? Python vs R ? Python on Apache Spark ? Python Dev Tools and REPLs ? Anaconda ? IPython ? Visual Studio Code ? Jupyter ? Jupyter Basic Commands ? Summary Applied Data Science ? What is Data Science? ? Data Science Ecosystem ? Data Mining vs. Data Science ? Business Analytics vs. Data Science ? Data Science, Machine Learning, AI? ? Who is a Data Scientist? ? Data Science Skill Sets Venn Diagram ? Data Scientists at Work ? Examples of Data Science Projects ? An Example of a Data Product ? Applied Data Science at Google ? Data Science Gotchas ? Summary Data Analytics Life-cycle Phases ? Big Data Analytics Pipeline ? Data Discovery Phase ? Data Harvesting Phase ? Data Priming Phase ? Data Logistics and Data Governance ? Exploratory Data Analysis ? Model Planning Phase ? Model Building Phase ? Communicating the Results ? Production Roll-out ? Summary Repairing and Normalizing Data ? Repairing and Normalizing Data ? Dealing with the Missing Data ? Sample Data Set ? Getting Info on Null Data ? Dropping a Column ? Interpolating Missing Data in pandas ? Replacing the Missing Values with the Mean Value ? Scaling (Normalizing) the Data ? Data Preprocessing with scikit-learn ? Scaling with the scale() Function ? The MinMaxScaler Object ? Summary Descriptive Statistics Computing Features in Python ? Descriptive Statistics ? Non-uniformity of a Probability Distribution ? Using NumPy for Calculating Descriptive Statistics Measures ? Finding Min and Max in NumPy ? Using pandas for Calculating Descriptive Statistics Measures ? Correlation ? Regression and Correlation ? Covariance ? Getting Pairwise Correlation and Covariance Measures ? Finding Min and Max in pandas DataFrame ? Summary Data Aggregation and Grouping ? Data Aggregation and Grouping ? Sample Data Set ? The pandas.core.groupby.SeriesGroupBy Object ? Grouping by Two or More Columns ? Emulating the SQL's WHERE Clause ? The Pivot Tables ? Cross-Tabulation ? Summary Data Visualization with matplotlib ? Data Visualization ? What is matplotlib? ? Getting Started with matplotlib ? The Plotting Window ? The Figure Options ? The matplotlib.pyplot.plot() Function ? The matplotlib.pyplot.bar() Function ? The matplotlib.pyplot.pie () Function ? Subplots ? Using the matplotlib.gridspec.GridSpec Object ? The matplotlib.pyplot.subplot() Function ? Hands-on Exercise ? Figures ? Saving Figures to File ? Visualization with pandas ? Working with matplotlib in Jupyter Notebooks ? Summary Data Science and ML Algorithms in scikit-learn ? Data Science, Machine Learning, AI? ? Types of Machine Learning ? Terminology: Features and Observations ? Continuous and Categorical Features (Variables) ? Terminology: Axis ? The scikit-learn Package ? scikit-learn Estimators ? Models, Estimators, and Predictors ? Common Distance Metrics ? The Euclidean Metric ? The LIBSVM format ? Scaling of the Features ? The Curse of Dimensionality ? Supervised vs Unsupervised Machine Learning ? Supervised Machine Learning Algorithms ? Unsupervised Machine Learning Algorithms ? Choose the Right Algorithm ? Life-cycles of Machine Learning Development ? Data Split for Training and Test Data Sets ? Data Splitting in scikit-learn ? Hands-on Exercise ? Classification Examples ? Classifying with k-Nearest Neighbors (SL) ? k-Nearest Neighbors Algorithm ? k-Nearest Neighbors Algorithm ? The Error Rate ? Hands-on Exercise ? Dimensionality Reduction ? The Advantages of Dimensionality Reduction ? Principal component analysis (PCA) ? Hands-on Exercise ? Data Blending ? Decision Trees (SL) ? Decision Tree Terminology ? Decision Tree Classification in Context of Information Theory ? Information Entropy Defined ? The Shannon Entropy Formula ? The Simplified Decision Tree Algorithm ? Using Decision Trees ? Random Forests ? SVM ? Naive Bayes Classifier (SL) ? Naive Bayesian Probabilistic Model in a Nutshell ? Bayes Formula ? Classification of Documents with Naive Bayes ? Unsupervised Learning Type: Clustering ? Clustering Examples ? k-Means Clustering (UL) ? k-Means Clustering in a Nutshell ? k-Means Characteristics ? Regression Analysis ? Simple Linear Regression Model ? Linear vs Non-Linear Regression ? Linear Regression Illustration ? Major Underlying Assumptions for Regression Analysis ? Least-Squares Method (LSM) ? Locally Weighted Linear Regression ? Regression Models in Excel ? Multiple Regression Analysis ? Logistic Regression ? Regression vs Classification ? Time-Series Analysis ? Decomposing Time-Series ? Summary Lab Exercises Lab 1 - Learning the Lab Environment Lab 2 - Using Jupyter Notebook Lab 3 - Repairing and Normalizing Data Lab 4 - Computing Descriptive Statistics Lab 5 - Data Grouping and Aggregation Lab 6 - Data Visualization with matplotlib Lab 7 - Data Splitting Lab 8 - k-Nearest Neighbors Algorithm Lab 9 - The k-means Algorithm Lab 10 - The Random Forest Algorithm

Python With Data Science
Delivered OnlineFlexible Dates
Price on Enquiry

Python for Data Analytics

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This course is aimed at anyone who wants to harness the power of data analytics in their organization including: Business Analysts, Data Analysts, Reporting and BI professionals Analytics professionals and Data Scientists who would like to learn Python Overview This course teaches delegates with no prior programming or data analytics experience how to perform data manipulation, data analysis and data visualization in Python. Mastery of these techniques and how to apply them to business problems will allow delegates to immediately add value in their workplace by extracting valuable insight from company data to allow better, data-driven decisions. Outcome: After attending this course, delegates will: Be able to write effective Python code Know how to access their data from a variety of sources using Python Know how to identify and fix data quality using Python Know how to manipulate data to create analysis ready data Know how to analyze and visualize data to drive data driven decisioning across your organization Becoming a world class data analytics practitioner requires mastery of the most sophisticated data analytics tools. These programming languages are some of the most powerful and flexible tools in the data analytics toolkit. From business questions to data analytics, and beyond For data analytics tasks to affect business decisions they must be driven by a business question. This section will formally outline how to move an analytics project through key phases of development from business question to business solution. Delegates will be able: to describe and understand the general analytics process. to describe and understand the different types of analytics can be used to derive data driven solutions to business to apply that knowledge to their business context Basic Python Programming Conventions This section will cover the basics of writing R programs. Topics covered will include: What is Python? Using Anaconda Writing Python programs Expressions and objects Functions and arguments Basic Python programming conventions Data Structures in Python This section will look at the basic data structures that Python uses and accessing data in Python. Topics covered will include: Vectors Arrays and matrices Factors Lists Data frames Loading .csv files into Python Connecting to External Data This section will look at loading data from other sources into Python. Topics covered will include: Loading .csv files into a pandas data frame Connecting to and loading data from a database into a panda data frame Data Manipulation in Python This section will look at how Python can be used to perform data manipulation operations to prepare datasets for analytics projects. Topics covered will include: Filtering data Deriving new fields Aggregating data Joining data sources Connecting to external data sources Descriptive Analytics and Basic Reporting in Python This section will explain how Python can be used to perform basic descriptive. Topics covered will include: Summary statistics Grouped summary statistics Using descriptive analytics to assess data quality Using descriptive analytics to created business report Using descriptive analytics to conduct exploratory analysis Statistical Analysis in Python This section will explain how Python can be used to created more interesting statistical analysis. Topics covered will include: Significance tests Correlation Linear regressions Using statistical output to create better business decisions. Data Visualisation in Python This section will explain how Python can be used to create effective charts and visualizations. Topics covered will include: Creating different chart types such as bar charts, box plots, histograms and line plots Formatting charts Best Practices Hints and Tips This section will go through some best practice considerations that should be adopted of you are applying Python in a business context.

Python for Data Analytics
Delivered OnlineFlexible Dates
Price on Enquiry

Effective Data Visualisation

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for This course is aimed at anyone currently working with data who is interested in using data visualisation to more effectively communicate their results. Overview At completion, delegates will understand how data visualisations can be best used to communicate actionable insights from data and be competent with the tools required to do it. Visualising data, and analytics results, is one of the most effective ways to achieve this. This course will cover the theory of data visualisation along with practical skills for creating compelling visualisations from data. Course Outline The use of analytics, statistics and data science in business has grown massively in recent years. Harnessing the power of data is opening actionable insights in diverse industries from banking to horse breeding. The companies doing this most successfully understand that using sophisticated analytics approaches to unlock insights from data is only half the job. Communicating these insights to all of the different parts of an organisation is just as important as doing the actual analysis. Visualising data, and analytics results, is one of the most effective ways to achieve this. This course will cover the theory of data visualisation along with practical skills for creating compelling visualisations from data. To attend this course delegates should be competent in the use of data analysis tools such as reporting tools, spreadsheet software or business intelligence tools. The course will explore the following topics through a series of interactive workshop sessions: Fundamentals of data visualisation Data characteristics & dimensions Mapping visual encodings to data dimensions Colour theory Graphical perception & communication Interaction design Visualisation different characteristics of data: trends, comparisons, correlations, maps, networks, hierarchies, text Designing effective dashboards

Effective Data Visualisation
Delivered OnlineFlexible Dates
Price on Enquiry

Beginning Data Analytics With R

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This course is aimed at anyone who wants to harness the power of data analytics in their organization. Overview After completing this course delegates will be capable of writing effective R code to manipulate, analyse and visualise data to enable their organisations make better, data-driven decisions. This course teaches delegates with no prior programming or data analytics experience how to perform data manipulation, data analysis and data visualisation in R. Course Outline Becoming a world class data analytics practitioner requires mastery of the most sophisticated data analytics tools. The R programming language is one of the most powerful and flexible tools in the data analytics toolkit. This course teaches delegates with no prior programming or data analytics experience how to perform data manipulation, data analysis and data visualisation in R. Mastery of these techniques will allow delegates to immediately add value in their work place by extracting valuable insight from company data to allow better, data-driven decisions. The course will explore the following topics through a series of interactive workshop sessions: What is R? Basic R programming conventions Data structures in R Accessing data in R Descriptive statistics in R Statistical analysis in R Data manipulation in R Data visualisation in R Additional course details: Nexus Humans Beginning Data Analytics With R training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Beginning Data Analytics With R course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Beginning Data Analytics With R
Delivered OnlineFlexible Dates
Price on Enquiry

Data Engineering on Google Cloud

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This class is intended for experienced developers who are responsible for managing big data transformations including: Extracting, loading, transforming, cleaning, and validating data. Designing pipelines and architectures for data processing. Creating and maintaining machine learning and statistical models. Querying datasets, visualizing query results and creating reports Overview Design and build data processing systems on Google Cloud Platform. Leverage unstructured data using Spark and ML APIs on Cloud Dataproc. Process batch and streaming data by implementing autoscaling data pipelines on Cloud Dataflow. Derive business insights from extremely large datasets using Google BigQuery. Train, evaluate and predict using machine learning models using TensorFlow and Cloud ML. Enable instant insights from streaming data Get hands-on experience with designing and building data processing systems on Google Cloud. This course uses lectures, demos, and hand-on labs to show you how to design data processing systems, build end-to-end data pipelines, analyze data, and implement machine learning. This course covers structured, unstructured, and streaming data. Introduction to Data Engineering Explore the role of a data engineer. Analyze data engineering challenges. Intro to BigQuery. Data Lakes and Data Warehouses. Demo: Federated Queries with BigQuery. Transactional Databases vs Data Warehouses. Website Demo: Finding PII in your dataset with DLP API. Partner effectively with other data teams. Manage data access and governance. Build production-ready pipelines. Review GCP customer case study. Lab: Analyzing Data with BigQuery. Building a Data Lake Introduction to Data Lakes. Data Storage and ETL options on GCP. Building a Data Lake using Cloud Storage. Optional Demo: Optimizing cost with Google Cloud Storage classes and Cloud Functions. Securing Cloud Storage. Storing All Sorts of Data Types. Video Demo: Running federated queries on Parquet and ORC files in BigQuery. Cloud SQL as a relational Data Lake. Lab: Loading Taxi Data into Cloud SQL. Building a Data Warehouse The modern data warehouse. Intro to BigQuery. Demo: Query TB+ of data in seconds. Getting Started. Loading Data. Video Demo: Querying Cloud SQL from BigQuery. Lab: Loading Data into BigQuery. Exploring Schemas. Demo: Exploring BigQuery Public Datasets with SQL using INFORMATION_SCHEMA. Schema Design. Nested and Repeated Fields. Demo: Nested and repeated fields in BigQuery. Lab: Working with JSON and Array data in BigQuery. Optimizing with Partitioning and Clustering. Demo: Partitioned and Clustered Tables in BigQuery. Preview: Transforming Batch and Streaming Data. Introduction to Building Batch Data Pipelines EL, ELT, ETL. Quality considerations. How to carry out operations in BigQuery. Demo: ELT to improve data quality in BigQuery. Shortcomings. ETL to solve data quality issues. Executing Spark on Cloud Dataproc The Hadoop ecosystem. Running Hadoop on Cloud Dataproc. GCS instead of HDFS. Optimizing Dataproc. Lab: Running Apache Spark jobs on Cloud Dataproc. Serverless Data Processing with Cloud Dataflow Cloud Dataflow. Why customers value Dataflow. Dataflow Pipelines. Lab: A Simple Dataflow Pipeline (Python/Java). Lab: MapReduce in Dataflow (Python/Java). Lab: Side Inputs (Python/Java). Dataflow Templates. Dataflow SQL. Manage Data Pipelines with Cloud Data Fusion and Cloud Composer Building Batch Data Pipelines visually with Cloud Data Fusion. Components. UI Overview. Building a Pipeline. Exploring Data using Wrangler. Lab: Building and executing a pipeline graph in Cloud Data Fusion. Orchestrating work between GCP services with Cloud Composer. Apache Airflow Environment. DAGs and Operators. Workflow Scheduling. Optional Long Demo: Event-triggered Loading of data with Cloud Composer, Cloud Functions, Cloud Storage, and BigQuery. Monitoring and Logging. Lab: An Introduction to Cloud Composer. Introduction to Processing Streaming Data Processing Streaming Data. Serverless Messaging with Cloud Pub/Sub Cloud Pub/Sub. Lab: Publish Streaming Data into Pub/Sub. Cloud Dataflow Streaming Features Cloud Dataflow Streaming Features. Lab: Streaming Data Pipelines. High-Throughput BigQuery and Bigtable Streaming Features BigQuery Streaming Features. Lab: Streaming Analytics and Dashboards. Cloud Bigtable. Lab: Streaming Data Pipelines into Bigtable. Advanced BigQuery Functionality and Performance Analytic Window Functions. Using With Clauses. GIS Functions. Demo: Mapping Fastest Growing Zip Codes with BigQuery GeoViz. Performance Considerations. Lab: Optimizing your BigQuery Queries for Performance. Optional Lab: Creating Date-Partitioned Tables in BigQuery. Introduction to Analytics and AI What is AI?. From Ad-hoc Data Analysis to Data Driven Decisions. Options for ML models on GCP. Prebuilt ML model APIs for Unstructured Data Unstructured Data is Hard. ML APIs for Enriching Data. Lab: Using the Natural Language API to Classify Unstructured Text. Big Data Analytics with Cloud AI Platform Notebooks What's a Notebook. BigQuery Magic and Ties to Pandas. Lab: BigQuery in Jupyter Labs on AI Platform. Production ML Pipelines with Kubeflow Ways to do ML on GCP. Kubeflow. AI Hub. Lab: Running AI models on Kubeflow. Custom Model building with SQL in BigQuery ML BigQuery ML for Quick Model Building. Demo: Train a model with BigQuery ML to predict NYC taxi fares. Supported Models. Lab Option 1: Predict Bike Trip Duration with a Regression Model in BQML. Lab Option 2: Movie Recommendations in BigQuery ML. Custom Model building with Cloud AutoML Why Auto ML? Auto ML Vision. Auto ML NLP. Auto ML Tables.

Data Engineering on Google Cloud
Delivered OnlineFlexible Dates
Price on Enquiry

DS10 SAP Data Services - Platform and Transforms

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for The primary audience for this course are Application Consultants, Developers, Developer Consultants, and Technology Consultants. Overview Define Data ServicesDefine Source and Target ConnectionsTrace, Validate, and Debug Data Services JobsUse Data Services TransformsImplement Change Data Capture in Data Services In this course, students will learn how to define data services, source, and target connections, as well as use data services transforms and implement change in data capture within data services. Data Services Defining Data Services Source and Target Metadata Defining Datastores in Data Services Defining a Data Services Flat File Format Batch Job Creation Creating Batch Jobs Batch Job Troubleshooting Writing Comments with Descriptions and Annotations Validating and Tracing Jobs Debugging Data Flows Auditing Data Flows Functions, Scripts, and Variables Using Built-In Functions Using Variables, Parameters, and Scripts Platform Transforms Using Platform Transforms Using the Map Operation Transform Using the Validation Transform Using the Merge Transform Using the Case Transform Using the SQL Transform Error Handling Setting Up Error Handling Changes in Data Capturing Changes in Data Using Source-Based Change Data Capture (CDC) Using Target-Based Change Data Capture (CDC) Data Services (Integrator) Platform Transforms Using Data Services (Integrator) Platform Transforms Using the Pivot Transform Using the Data Transfer Transform Additional course details: Nexus Humans DS10 SAP Data Services - Platform and Transforms training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the DS10 SAP Data Services - Platform and Transforms course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

DS10 SAP Data Services - Platform and Transforms
Delivered OnlineFlexible Dates
Price on Enquiry

Data-driven Business Using Statistical Analysis

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for This course is suited to marketeers, business analysts, and researchers who are interested in increasing their statistical knowledge. Overview After attending this course, delegates will understand how statistics can be used to provide valuable insight into their business, and be able to apply statistical methods to solve business problems. On returning to work delegates will immediately be able to make a difference to the way that their organisations make decisions. This course covers the statistical methods that analysts need to move from simple reporting on business problems to extracting insight to solve business problems. Course Outline The course will explore the following topics through a series of lectures and workshops: Summary statistics for both continuous data and categorical data Using and reporting confidence intervals Using hypothesis tests to answer business questions Using correlations to explore data relationships Simple prediction models Analysing categorical data Additional course details: Nexus Humans Data-driven Business Using Statistical Analysis training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Data-driven Business Using Statistical Analysis course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Data-driven Business Using Statistical Analysis
Delivered OnlineFlexible Dates
Price on Enquiry

Writing Analytical Queries for Business Intelligence

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This course is intended for information workers and data science professionals who seek to use database reporting and analysis tools such as Microsoft SQL Server Reporting Services, Excel, Power BI, R, SAS and other business intelligence tools, and wish to use TSQL queries to efficiently retrieve data sets from Microsoft SQL Server relational databases for use with these tools. Overview Identify independent and dependent variables and measurement levels in their own analytical work scenarios. Identify variables of interest in relational database tables. Choose a data aggregation level and data set design appropriate for the intended analysis and tool. Use TSQL SELECT queries to produce ready-to-use data sets for analysis in tools such as PowerBI, SQL Server Reporting Services, Excel, R, SAS, SPSS, and others. Create stored procedures, views, and functions to modularize data retrieval code. This three-day instructor-led course is about writing TSQL queries for the purpose of database reporting, analysis, and business intelligence. Specifically, this course presents TSQL within the context of data analysis in other words, making meaning from the data rather than transaction-oriented data-tier application development. The course starts with a brief discussion of levels of measurement and quantitative research methodogy, and integrates these concepts into each TSQL topic presented. The goal is to provide a consistent, direct, and purposeful learning path from RDBMS data retrieval through analytical tools such as SQL Server Reporting Services, PowerBI, Excel, R, SAS, and SPSS. Module 1: Introduction to TSQL for Business Intelligence Two Approaches to SQL Programming TSQL Data Retrieval in an Analytics / Business Intelligence Environment The Database Engine SQL Server Management Studio and the CarDeal Sample Database Identifying Variables in Tables SQL is a Declarative Language Introduction to the SELECT Query Module 2: Turning Table Columns into Variables for Analysis: SELECT List Expressions, WHERE, and ORDER BY Turning Columns into Variables for Analysis Column Expressions, Data Types, and Built-in Functions Column aliases Data type conversions Built-in Scalar Functions Table Aliases The WHERE clause ORDER BY Module 3: Combining Columns from Multiple Tables into a Single Dataset: The JOIN Operators Primary Keys, Foreign Keys, and Joins Understanding Joins, Part 1: CROSSJOIN and the Full Cartesian Product Understanding Joins, Part 2: The INNERJOIN Understanding Joins, Part 3: The OUTERJOINS Understanding Joins, Part 4: Joining more than two tables Understanding Joins, Part 5: Combining INNER and OUTERJOINs Combining JOIN Operations with WHERE and ORDER BY Module 4: Creating an Appropriate Aggregation Level Using GROUP BY Identifying required aggregation level and granularity Aggregate Functions GROUP BY HAVING Order of operations in SELECT queries Module 5: Subqueries, Derived Tables and Common Table Expressions Non-correlated and correlated subqueries Derived tables Common table expressions Module 6: Encapsulating Data Retrieval Logic Views Table-valued functions Stored procedures Creating objects for read-access users Creating database accounts for analytical client tools Module 7: Getting Your Dataset to the Client Connecting to SQL Server and Submitting Queries from Client Tools Connecting and running SELECT queries from: Excel PowerBI RStudio Exporting datasets to files using Results pane from SSMS The bcp utility The Import/Export Wizard Additional course details: Nexus Humans Writing Analytical Queries for Business Intelligence training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Writing Analytical Queries for Business Intelligence course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Writing Analytical Queries for Business Intelligence
Delivered OnlineFlexible Dates
Price on Enquiry

KM213 IBM InfoSphere QualityStage Essentials v11.5

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for Data Analysts responsible for data quality using QualityStageData Quality ArchitectsData Cleansing Developers Overview List the common data quality contaminantsDescribe each of the following processes: Investigation, Standardization, Match. & SurvivorshipDescribe QualityStage architectureDescribe QualityStage clients and their functionsImport metadataBuild and run DataStage/QualityStage jobs, review resultsBuild Investigate jobsUse Character Discrete, Concatenate, and Word Investigations to analyze data fieldsDescribe the Standardize stageIdentify Rule SetsBuild jobs using the Standardize stageInterpret standardization resultsInvestigate unhandled data and patternsBuild a QualityStage job to identify matching recordsApply multiple Match passes to increase efficiencyInterpret and improve match resultsBuild a QualityStage Survive job that will consolidate matched records into a single master recordBuild a single job to match data using a Two-Source match This course teaches how to build QualityStage parallel jobs that investigate, standardize, match, and consolidate data records. Students will gain experience by building an application that combines customer data from three source systems. Data Quality Issues Listing the common data quality contaminants Describing data quality processes QualityStage Overview Describing QualityStage architecture Describing QualityStage clients and their functions Developing with QualityStage Importing metadata Building DataStage/QualityStage Jobs Running jobs Reviewing results Investigate Building Investigate jobs Using Character Discrete, Concatenate, and Word Investigations to analyze data fields Reviewing results Standardize Describing the Standardize stage Identifying Rule Sets Building jobs using the Standardize stage Interpreting standardize results Investigating unhandled data and patterns Match Building a QualityStage job to identify matching records Applying multiple Match passes to increase efficiency Interpreting and improving Match results Survive Building a QualityStage survive job that will consolidate matched records into a single master record Two-Source Match Building a QualityStage job to match data using a reference match Additional course details: Nexus Humans KM213 IBM InfoSphere QualityStage Essentials v11.5 training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the KM213 IBM InfoSphere QualityStage Essentials v11.5 course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

KM213 IBM InfoSphere QualityStage Essentials v11.5
Delivered OnlineFlexible Dates
Price on Enquiry