Duration 3 Days 18 CPD hours This course is intended for If you have worked in C++ but want to learn how to make the most of this language, especially for large projects, this course is for you. Overview By the end of this course, you'll have developed programming skills that will set you apart from other C++ programmers. After completing this course, you will be able to: Delve into the anatomy and workflow of C++ Study the pros and cons of different approaches to coding in C++ Test, run, and debug your programs Link object files as a dynamic library Use templates, SFINAE, constexpr if expressions and variadic templates Apply best practice to resource management This course begins with advanced C++ concepts by helping you decipher the sophisticated C++ type system and understand how various stages of compilation convert source code to object code. You'll then learn how to recognize the tools that need to be used in order to control the flow of execution, capture data, and pass data around. By creating small models, you'll even discover how to use advanced lambdas and captures and express common API design patterns in C++. As you cover later lessons, you'll explore ways to optimize your code by learning about memory alignment, cache access, and the time a program takes to run. The concluding lesson will help you to maximize performance by understanding modern CPU branch prediction and how to make your code cache-friendly. Anatomy of Portable C++ Software Managing C++ Projects Writing Readable Code No Ducks Allowed ? Types and Deduction C++ Types Creating User Types Structuring our Code No Ducks Allowed ? Templates and Deduction Inheritance, Polymorphism, and Interfaces Templates ? Generic Programming Type Aliases ? typedef and using Class Templates No Leaks Allowed ? Exceptions and Resources Exceptions in C++ RAII and the STL Move Semantics Name Lookup Caveat Emptor Separation of Concerns ? Software Architecture, Functions, and Variadic Templates Function Objects and Lambda Expressions Variadic Templates The Philosophers' Dinner ? Threads and Concurrency Synchronous, Asynchronous, and Threaded Execution Review Synchronization, Data Hazards, and Race Conditions Future, Promises, and Async Streams and I/O File I/O Implementation Classes String I/O Implementation I/O Manipulators Making Additional Streams Using Macros Everybody Falls, It's How You Get Back Up ? Testing and Debugging Assertions Unit Testing and Mock Testing Understanding Exception Handling Breakpoints, Watchpoints, and Data Visualization Need for Speed ? Performance and Optimization Performance Measurement Runtime Profiling Optimization Strategies Cache Friendly Code
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Architecture and course review Post course knowledge check Architecture review Course review
Overview Data and visual analytics are emerging fields concerned with analysing, modelling, and visualizing complex high-dimensional data. It can be analysed and visualised with many languages like Python, R Programming and more. This course will help to attain the skills and give in-depth knowledge to the participant's enhanced way of modelling, analysing and visualizing techniques. The course will highlight practical challenges including composite real-world data and will also comprise several practical studies
Duration 5 Days 30 CPD hours This course is intended for This course is designed for students who want to learn the R programming language, particularly students who want to leverage R for data analysis and data science tasks in their organization. The course is also designed for students with an interest in applying statistics to real-world problems. A typical student in this course should have several years of experience with computing technology, along with a proficiency in at least one other programming language. Overview In this course, you will use R to perform common data science tasks.You will: Set up an R development environment and execute simple code. Perform operations on atomic data types in R, including characters, numbers, and logicals. Perform operations on data structures in R, including vectors, lists, and data frames. Write conditional statements and loops. Structure code for reuse with functions and packages. Manage data by loading and saving datasets, manipulating data frames, and more. Analyze data through exploratory analysis, statistical analysis, and more. Create and format data visualizations using base R and ggplot2. Create simple statistical models from data. In our data-driven world, organizations need the right tools to extract valuable insights from that data. The R programming language is one of the tools at the forefront of data science. Its robust set of packages and statistical functions makes it a powerful choice for analyzing data, manipulating data, performing statistical tests on data, and creating predictive models from data. Likewise, R is notable for its strong data visualization tools, enabling you to create high-quality graphs and plots that are incredibly customizable. This course will teach you the fundamentals of programming in R to get you started. It will also teach you how to use R to perform common data science tasks and achieve data-driven results for the business. Lesson 1: Setting Up R and Executing Simple Code Topic A: Set Up the R Development Environment Topic B: Write R Statements Lesson 2: Processing Atomic Data Types Topic A: Process Characters Topic B: Process Numbers Topic C: Process Logicals Lesson 3: Processing Data Structures Topic A: Process Vectors Topic B: Process Factors Topic C: Process Data Frames Topic D: Subset Data Structures Lesson 4: Writing Conditional Statements and Loops Topic A: Write Conditional Statements Topic B: Write Loops Lesson 5: Structuring Code for Reuse Topic A: Define and Call Functions Topic B: Apply Loop Functions Topic C: Manage R Packages Lesson 6: Managing Data in R Topic A: Load Data Topic B: Save Data Topic C: Manipulate Data Frames Using Base R Topic D: Manipulate Data Frames Using dplyr Topic E: Handle Dates and Times Lesson 7: Analyzing Data in R Topic A: Examine Data Topic B: Explore the Underlying Distribution of Data Topic C: Identify Missing Values Lesson 8: Visualizing Data in R Topic A: Plot Data Using Base R Functions Topic B: Plot Data Using ggplot2 Topic C: Format Plots in ggplot2 Topic D: Create Combination Plots Lesson 9: Modeling Data in R Topic A: Create Statistical Models in R Topic B: Create Machine Learning Models in R
Data dashboards provide key information to stakeholders so that they can make informed decisions. While there are plenty of software solutions for building these essential data products, there is much less guidance on how to design dashboards to meet the diverse needs of users. This course is for anyone who is building or implementing dashboards, and wants to know more about design principles and best practice. You could be using business intelligence software (such as Power BI or Tableau), or implementing bespoke solutions. The course will give your team the ability to evaluate user needs and levels of understanding, make informed decisions about chart selections, and make effective use of interactivity dynamic data. We’ll work with you before the course to ensure that we understand your organisation and what you’re hoping to achieve. Sample learning content Session 1: Data with a purpose Understanding the different types of dashboard. Information overload and other common dashboard pitfalls. Assessing user needs and levels of data fluency. Session 2: Planning a dashboard Assessing diverse user needs and levels of data fluency. Taking a User Experience (UX) approach to design and navigation. Applying an interative and collaborative approach to onboarding. Session 3: Graphs, charts and dials Understanding how graphical perception informs chart choices. Making intelligent design choices to help users explore. Design principles for layout and navigation. Session 4: Using interactivity Making effective use of filters to slice and dice data sets. Using layers of information to enable drilldown data exploration. Complenting dashboards with automated alerts and queries. Delivery We deliver our courses over Zoom, to maximise flexibility. The training can be delivered in a single day, or across multiple sessions. All of our courses are live and interactive – every session includes a mix of formal tuition and hands-on exercises. To ensure this is possible, the number of attendees is capped at 16 people. Tutor Alan Rutter is the founder of Fire Plus Algebra. He is a specialist in communicating complex subjects through data visualisation, writing and design. He teaches for General Assembly and runs in-house training for public sector clients including the Home Office, the Department of Transport, the Biotechnology and Biological Sciences Research Council, the Health Foundation, and numerous local government and emergency services teams. He previously worked with Guardian Masterclasses on curating and delivering new course strands, including developing and teaching their B2B data visualisation courses. He oversaw the iPad edition launches of Wired, GQ, Vanity Fair and Vogue in the UK, and has worked with Condé Nast International as product owner on a bespoke digital asset management system for their 11 global markets. Testimonial “Alan was great to work with, he took us through the concepts behind data visualisation which means our team is now equipped for the future. He has a wide range of experience across the topic that is delivered in a clear, concise and friendly manner. We look forward to working with Alan again in the future.” John Masterson | Chief Product Officer | ImproveWell
Duration 1 Days 6 CPD hours This course is intended for This course is intended for: Data platform engineers Solutions architects IT professionals Overview In this course, you will learn to: Apply data lake methodologies in planning and designing a data lake Articulate the components and services required for building an AWS data lake Secure a data lake with appropriate permission Ingest, store, and transform data in a data lake Query, analyze, and visualize data within a data lake In this course, you will learn how to build an operational data lake that supports analysis of both structured and unstructured data. You will learn the components and functionality of the services involved in creating a data lake. You will use AWS Lake Formation to build a data lake, AWS Glue to build a data catalog, and Amazon Athena to analyze data. The course lectures and labs further your learning with the exploration of several common data lake architectures. Module 1: Introduction to data lakes Describe the value of data lakes Compare data lakes and data warehouses Describe the components of a data lake Recognize common architectures built on data lakes Module 2: Data ingestion, cataloging, and preparation Describe the relationship between data lake storage and data ingestion Describe AWS Glue crawlers and how they are used to create a data catalog Identify data formatting, partitioning, and compression for efficient storage and query Lab 1: Set up a simple data lake Module 3: Data processing and analytics Recognize how data processing applies to a data lake Use AWS Glue to process data within a data lake Describe how to use Amazon Athena to analyze data in a data lake Module 4: Building a data lake with AWS Lake Formation Describe the features and benefits of AWS Lake Formation Use AWS Lake Formation to create a data lake Understand the AWS Lake Formation security model Lab 2: Build a data lake using AWS Lake Formation Module 5: Additional Lake Formation configurations Automate AWS Lake Formation using blueprints and workflows Apply security and access controls to AWS Lake Formation Match records with AWS Lake Formation FindMatches Visualize data with Amazon QuickSight Lab 3: Automate data lake creation using AWS Lake Formation blueprints Lab 4: Data visualization using Amazon QuickSight Module 6: Architecture and course review Post course knowledge check Architecture review Course review Additional course details: Nexus Humans Building Data Lakes on AWS training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Building Data Lakes on AWS course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.
Duration 2 Days 12 CPD hours Overview Identify and configure basic functions of Tableau. Connect to data sources, import data into Tableau, and save Tableau files Create views and customize data in visualizations. Manage, sort, and group data. Save and share data sources and workbooks. Filter data in views. Customize visualizations with annotations, highlights, and advanced features. Create and enhance dashboards in Tableau. Create and enhance stories in Tableau As technology progresses and becomes more interwoven with our businesses and lives, more and more data is collected about business and personal activities. This era of "big data" has exploded due to the rise of cloud computing, which provides an abundance of computational power and storage, allowing organizations of all sorts to capture and store data. Leveraging that data effectively can provide timely insights and competitive advantage. The creation of data-backed visualizations is a key way data scientists, or any professional, can explore, analyze, and report insights and trends from data. Tableau© software is designed for this purpose. Tableau was built to connect to a wide range of data sources and allows users to quickly create visualizations of connected data to gain insights, show trends, and create reports. Tableau's data connection capabilities and visualization features go far beyond those that can be found in spreadsheets, allowing users to create compelling and interactive worksheets, dashboards, and stories that bring data to life and turn data into thoughtful action. Prerequisites To ensure your success in this course, you should have experience managing data with Microsoft© Excel© or Google Sheets?. Lesson 1: Tableau Fundamentals Topic A: Overview of Tableau Topic B: Navigate and Configure Tableau Lesson 2: Connecting to and Preparing Data Topic A: Connect to Data Topic B: Build a Data Model Topic C: Save Workbook Files Topic D: Prepare Data for Analysis Lesson 3: Exploring Data Topic A: Create Views Topic B: Customize Data in Visualizations Lesson 4: Managing, Sorting, and Grouping Data Topic A: Adjust Fields Topic B: Sort Data Topic C: Group Data Lesson 5: Saving, Publishing, and Sharing Data Topic A: Save Data Sources Topic B: Publish Data Sources and Visualizations Topic C: Share Workbooks for Collaboration Lesson 6: Filtering Data Topic A: Configure Worksheet Filters Topic B: Apply Advanced Filter Options Topic C: Create Interactive Filters Lesson 7: Customizing Visualizations Topic A: Format and Annotate Views Topic B: Emphasize Data in Visualizations Topic C: Create Animated Workbooks Topic D: Best Practices for Visual Design Lesson 8: Creating Dashboards in Tableau Topic A: Create Dashboards Topic B: Enhance Dashboards with Actions Topic C: Create Mobile Dashboards Lesson 9: Creating Stories in Tableau Topic A: Create Stories Topic B: Enhance Stories with Tooltips
Duration 3 Days 18 CPD hours This course is intended for This course is aimed at anyone who wants to harness the power of data analytics in their organization including: Business Analysts, Data Analysts, Reporting and BI professionals Analytics professionals and Data Scientists who would like to learn Python Overview This course teaches delegates with no prior programming or data analytics experience how to perform data manipulation, data analysis and data visualization in Python. Mastery of these techniques and how to apply them to business problems will allow delegates to immediately add value in their workplace by extracting valuable insight from company data to allow better, data-driven decisions. Outcome: After attending this course, delegates will: Be able to write effective Python code Know how to access their data from a variety of sources using Python Know how to identify and fix data quality using Python Know how to manipulate data to create analysis ready data Know how to analyze and visualize data to drive data driven decisioning across your organization Becoming a world class data analytics practitioner requires mastery of the most sophisticated data analytics tools. These programming languages are some of the most powerful and flexible tools in the data analytics toolkit. From business questions to data analytics, and beyond For data analytics tasks to affect business decisions they must be driven by a business question. This section will formally outline how to move an analytics project through key phases of development from business question to business solution. Delegates will be able: to describe and understand the general analytics process. to describe and understand the different types of analytics can be used to derive data driven solutions to business to apply that knowledge to their business context Basic Python Programming Conventions This section will cover the basics of writing R programs. Topics covered will include: What is Python? Using Anaconda Writing Python programs Expressions and objects Functions and arguments Basic Python programming conventions Data Structures in Python This section will look at the basic data structures that Python uses and accessing data in Python. Topics covered will include: Vectors Arrays and matrices Factors Lists Data frames Loading .csv files into Python Connecting to External Data This section will look at loading data from other sources into Python. Topics covered will include: Loading .csv files into a pandas data frame Connecting to and loading data from a database into a panda data frame Data Manipulation in Python This section will look at how Python can be used to perform data manipulation operations to prepare datasets for analytics projects. Topics covered will include: Filtering data Deriving new fields Aggregating data Joining data sources Connecting to external data sources Descriptive Analytics and Basic Reporting in Python This section will explain how Python can be used to perform basic descriptive. Topics covered will include: Summary statistics Grouped summary statistics Using descriptive analytics to assess data quality Using descriptive analytics to created business report Using descriptive analytics to conduct exploratory analysis Statistical Analysis in Python This section will explain how Python can be used to created more interesting statistical analysis. Topics covered will include: Significance tests Correlation Linear regressions Using statistical output to create better business decisions. Data Visualisation in Python This section will explain how Python can be used to create effective charts and visualizations. Topics covered will include: Creating different chart types such as bar charts, box plots, histograms and line plots Formatting charts Best Practices Hints and Tips This section will go through some best practice considerations that should be adopted of you are applying Python in a business context.
Duration 1 Days 6 CPD hours This course is intended for The audience for this course includes professionals who are new to Looker who are interested in leveraging Looker for data analysis, visualization, and reporting. The course is designed for individuals seeking to gain a comprehensive understanding of Looker's functionalities and apply these skills in their organizations to drive data-driven decision-making. Overview Working in a hands-on learning environment led by our expert facilitator, you'll explore and gain: Solid foundation in Looker's platform: Acquire a comprehensive understanding of Looker's key features, functionality, and interface, enabling you to effectively utilize the platform for your data analysis and visualization needs. Proficiency in LookML and data modeling: Develop essential skills in Looker's unique data modeling language, LookML, to create efficient and customized data models tailored to your organization's specific requirements. Expertise in creating Explores: Learn how to build, customize, and save Explores with dimensions, measures, filters, and calculated fields, empowering you to analyze your data and uncover valuable insights in a short amount of time. Mastery of dashboard design and visualization: Gain the skills to design visually appealing and informative dashboards, create various types of visualizations, and customize them to effectively communicate your data story. Improved content organization with folders and boards: Understand how to effectively use folders and boards in Looker to organize, manage, and discover content, making your data insights easily accessible for you and your team. Looker Basics: Quick Start to Analyzing and Visualizing Data using Looker is a one day, hands-on course designed to equip professionals from a variety of backgrounds with the knowledge and skills needed to harness the full potential of their data using Looker's powerful platform. With the guidance of our expert trainers, you will gain a basic understanding of Looker's features, enabling you to create visually engaging, interactive, and insightful reports and dashboards to drive informed decision-making. Throughout this interactive workshop, you will explore Looker's key functionalities, including connecting to data sources, mastering LookML, building custom Explores, and designing captivating dashboards. With about 40% of the course dedicated to hands-on labs and a guided project, you will have ample opportunity to apply the skills you've learned in real world scenarios. Don't miss this opportunity to elevate your data analysis and visualization capabilities, enhance your professional skill set, and unlock the power of data-driven decision making. Getting Started with Looker Overview of Looker and its key features Navigating the Looker interface Connecting to Data Sources and LookML Basics Setting up and managing data connections Exploring database schemas Understanding LookML: Looker's data modeling language Creating and Customizing Explores Building and customizing Explores Adding dimensions, measures, and filters Creating calculated fields Data Visualization and Dashboard Design Creating visualizations using Looker's visualization library Customizing chart types, colors, and labels Displaying visualizations in dashboards Organizing Content with Folders and Boards Introduction to folders and boards in Looker Creating and managing folders for organizing content Setting up boards for easy content discovery Hands-on Workshop and Project Participants work on a guided project to apply the skills learned Wrap-up and Q&A Additional course details: Nexus Humans Looker Basics: Quick Start to Analyzing and Visualizing Data using Looker (TTDVLK01) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Looker Basics: Quick Start to Analyzing and Visualizing Data using Looker (TTDVLK01) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.