• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

551 Machine Learning (ML) courses delivered Online

Quick Start to Using Azure AI for Technical Users (TTAI2330)

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This introductory-level course is great for experienced technical professionals working in a wide range of industries, such as software development, data science, marketing and advertising, finance, healthcare, and more, who are looking to use the latest AI and machine learning techniques in their day to day. The hands-on labs in this course use Python, so you should have some familiarity with Python scripting basics. Overview Working in an interactive learning environment, led by our engaging OpenAI expert you'll: Understand the capabilities and products offered by OpenAI and how to access them through the OpenAI API. set up an OpenAI environment on Azure, including creating an Azure virtual machine and configuring the environment to connect to Azure resources. Gain hands-on experience building a GPT-3 based chatbot on Azure and implement advanced natural language processing capabilities. Use the OpenAI API to access GPT-3 and generate high-quality text Learn how to use Whisper to improve the quality of text generation. Understand the capabilities of DALL-E and use it to generate images for unique and engaging visuals. Geared for technical professionals, Quick Start to Azure AI Basics for Technical Users is a fun, fast paced course designed to quickly get you up to speed with OpenAI?s powerful tools and functionality, and to provide hands-on experience in setting up an OpenAI environment on Azure. Guided by our AI expert, you?ll explore the capabilities of OpenAI's GPT-3, Whisper and DALL-E, and build a chatbot on Azure. It will provide you with the knowledge and resources to continue your journey in AI and machine learning and have a good understanding of the potential of OpenAI and Azure for your projects. First, you?ll dive into the world of OpenAI, learning about its products and the capabilities they offer. You'll also discover how Azure's offerings for AI and machine learning can complement OpenAI's tools and resources, providing you with a powerful combination for your projects. And don't worry if you're new to Azure, we'll walk you through the process of setting up an account and creating a resource group. As you progress through the course, you'll get the chance to work with OpenAI's GPT-3, one of the most advanced large language models available today. You'll learn how to use the OpenAI API to access GPT-3 and discover how to use it to generate high-quality text quickly and easily. And that's not all, you'll also learn how to build a GPT-3 based chatbot on Azure, giving you the opportunity to implement advanced natural language processing capabilities in your chatbot projects. The course will also cover OpenAI Whisper, an OpenAI tool that can improve the quality of text generation, allowing you to create more coherent and natural language content. And you will learn about OpenAI DALL-E, an OpenAI tool that can generate images, giving you the ability to create unique and engaging visuals to enhance your content and projects. Introduction to OpenAI and Azure Explore OpenAI and its products, as well as Azure's offerings for AI and Machine Learning, allowing you to understand the tools and resources available to you for your AI projects. Explore OpenAI and its products Explore Azure and its offerings for AI and Machine Learning Get Hands-On: Setting up an OpenAI environment on Azure Walk through the process of setting up an OpenAI environment on Azure, giving you the hands-on experience needed to start building your own projects using OpenAI and Azure. Create an Azure virtual machine and installing the OpenAI SDK Configure the OpenAI environment and connecting to Azure resources Explore OpenAI GPT-3 Learn about GPT-3, one of OpenAI's most powerful language models, and how to use it to generate high quality text, giving you the ability to create natural language content quickly and easily. Review GPT-3 and its capabilities Use the OpenAI API to access GPT-3 Get Hands-on: Building a GPT-3 based chatbot on Azure Learn how to build a GPT-3 based chatbot on Azure, giving you the opportunity to learn how to implement advanced natural language processing capabilities in your chatbot projects. Setup an Azure Function and creating a chatbot Integrate GPT-3 with the chatbot OpenAI Whisper Explore Whisper, an OpenAI tool that can improve the quality of text generation, allowing you to create more coherent and natural language content. Explore Whisper and its capabilities Use Whisper to improve the quality of text generation OpenAI DALL-E Explore DALL-E, an OpenAI tool that can generate images, giving you the ability to create unique and engaging visuals to enhance your content and projects. Explore DALL-E and its capabilities Use the OpenAI API to access DALL-E What?s Next: Keep Going! Other ways OpenAI can impact your day to day Explore great places to check for expanded tools and add-ons for Azure OpenAI Where to go for help and support Quick Look at Generative AI and its Business Implications Understanding Generative AI Generative AI in Business Ethical considerations of Generative AI

Quick Start to Using Azure AI for Technical Users (TTAI2330)
Delivered OnlineFlexible Dates
Price on Enquiry

Data Science Projects with Python

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for If you are a data analyst, data scientist, or a business analyst who wants to get started with using Python and machine learning techniques to analyze data and predict outcomes, this book is for you. Basic knowledge of computer programming and data analytics is a must. Familiarity with mathematical concepts such as algebra and basic statistics will be useful. Overview By the end of this course, you will have the skills you need to confidently use various machine learning algorithms to perform detailed data analysis and extract meaningful insights from data. This course is designed to give you practical guidance on industry-standard data analysis and machine learning tools in Python, with the help of realistic data. The course will help you understand how you can use pandas and Matplotlib to critically examine a dataset with summary statistics and graphs, and extract the insights you seek to derive. You will continue to build on your knowledge as you learn how to prepare data and feed it to machine learning algorithms, such as regularized logistic regression and random forest, using the scikit-learn package. You?ll discover how to tune the algorithms to provide the best predictions on new and unseen data. As you delve into later sections, you?ll be able to understand the working and output of these algorithms and gain insight into not only the predictive capabilities of the models but also their reasons for making these predictions. Data Exploration and Cleaning Python and the Anaconda Package Management System Different Types of Data Science Problems Loading the Case Study Data with Jupyter and pandas Data Quality Assurance and Exploration Exploring the Financial History Features in the Dataset Activity 1: Exploring Remaining Financial Features in the Dataset Introduction to Scikit-Learn and Model Evaluation Introduction Model Performance Metrics for Binary Classification Activity 2: Performing Logistic Regression with a New Feature and Creating a Precision-Recall Curve Details of Logistic Regression and Feature Exploration Introduction Examining the Relationships between Features and the Response Univariate Feature Selection: What It Does and Doesn't Do Building Cloud-Native Applications Activity 3: Fitting a Logistic Regression Model and Directly Using the Coefficients The Bias-Variance Trade-off Introduction Estimating the Coefficients and Intercepts of Logistic Regression Cross Validation: Choosing the Regularization Parameter and Other Hyperparameters Activity 4: Cross-Validation and Feature Engineering with the Case Study Data Decision Trees and Random Forests Introduction Decision trees Random Forests: Ensembles of Decision Trees Activity 5: Cross-Validation Grid Search with Random Forest Imputation of Missing Data, Financial Analysis, and Delivery to Client Introduction Review of Modeling Results Dealing with Missing Data: Imputation Strategies Activity 6: Deriving Financial Insights Final Thoughts on Delivering the Predictive Model to the Client

Data Science Projects with Python
Delivered OnlineFlexible Dates
Price on Enquiry

Preparing for the Professional Data Engineer Examination

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for the following participants:Cloud professionals interested in taking the Data Engineer certification exam.Data engineering professionals interested in taking the Data Engineer certification exam. Overview This course teaches participants the following skills: Position the Professional Data Engineer Certification Provide information, tips, and advice on taking the exam Review the sample case studies Review each section of the exam covering highest-level concepts sufficient to build confidence in what is known by the candidate and indicate skill gaps/areas of study if not known by the candidate Connect candidates to appropriate target learning This course will help prospective candidates plan their preparation for the Professional Data Engineer exam. The session will cover the structure and format of the examination, as well as its relationship to other Google Cloud certifications. Through lectures, quizzes, and discussions, candidates will familiarize themselves with the domain covered by the examination, to help them devise a preparation strategy. Rehearse useful skills including exam question reasoning and case comprehension. Tips and review of topics from the Data Engineering curriculum. Understanding the Professional Data Engineer Certification Position the Professional Data Engineer certification among the offerings Distinguish between Associate and Professional Provide guidance between Professional Data Engineer and Associate Cloud Engineer Describe how the exam is administered and the exam rules Provide general advice about taking the exam Sample Case Studies for the Professional Data Engineer Exam Flowlogistic MJTelco Designing and Building (Review and preparation tips) Designing data processing systems Designing flexible data representations Designing data pipelines Designing data processing infrastructure Build and maintain data structures and databases Building and maintaining flexible data representations Building and maintaining pipelines Building and maintaining processing infrastructure Analyzing and Modeling (Review and preparation tips) Analyze data and enable machine learning Analyzing data Machine learning Machine learning model deployment Model business processes for analysis and optimization Mapping business requirements to data representations Optimizing data representations, data infrastructure performance and cost Reliability, Policy, and Security (Review and preparation tips) Design for reliability Performing quality control Assessing, troubleshooting, and improving data representation and data processing infrastructure Recovering data Visualize data and advocate policy Building (or selecting) data visualization and reporting tools Advocating policies and publishing data and reports Design for security and compliance Designing secure data infrastructure and processes Designing for legal compliance Resources and next steps Resources for learning more about designing data processing systems, data structures, and databases Resources for learning more about data analysis, machine learning, business process analysis, and optimization Resources for learning more about data visualization and policy Resources for learning more about reliability design Resources for learning more about business process analysis and optimization Resources for learning more about reliability, policies, security, and compliance Additional course details: Nexus Humans Preparing for the Professional Data Engineer Examination training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Preparing for the Professional Data Engineer Examination course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Preparing for the Professional Data Engineer Examination
Delivered OnlineFlexible Dates
Price on Enquiry

NLP Boot Camp / Hands-On Natural Language Processing (TTAI3030)

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This in an intermediate and beyond-level course is geared for experienced Python developers looking to delve into the exciting field of Natural Language Processing. It is ideally suited for roles such as data analysts, data scientists, machine learning engineers, or anyone working with text data and seeking to extract valuable insights from it. If you're in a role where you're tasked with analyzing customer sentiment, building chatbots, or dealing with large volumes of text data, this course will provide you with practical, hands on skills that you can apply right away. Overview This course combines engaging instructor-led presentations and useful demonstrations with valuable hands-on labs and engaging group activities. Throughout the course you'll: Master the fundamentals of Natural Language Processing (NLP) and understand how it can help in making sense of text data for valuable insights. Develop the ability to transform raw text into a structured format that machines can understand and analyze. Discover how to collect data from the web and navigate through semi-structured data, opening up a wealth of data sources for your projects. Learn how to implement sentiment analysis and topic modeling to extract meaning from text data and identify trends. Gain proficiency in applying machine learning and deep learning techniques to text data for tasks such as classification and prediction. Learn to analyze text sentiment, train emotion detectors, and interpret the results, providing a way to gauge public opinion or understand customer feedback. The Hands-on Natural Language Processing (NLP) Boot Camp is an immersive, three-day course that serves as your guide to building machines that can read and interpret human language. NLP is a unique interdisciplinary field, blending computational linguistics with artificial intelligence to help machines understand, interpret, and generate human language. In an increasingly data-driven world, NLP skills provide a competitive edge, enabling the development of sophisticated projects such as voice assistants, text analyzers, chatbots, and so much more. Our comprehensive curriculum covers a broad spectrum of NLP topics. Beginning with an introduction to NLP and feature extraction, the course moves to the hands-on development of text classifiers, exploration of web scraping and APIs, before delving into topic modeling, vector representations, text manipulation, and sentiment analysis. Half of your time is dedicated to hands-on labs, where you'll experience the practical application of your knowledge, from creating pipelines and text classifiers to web scraping and analyzing sentiment. These labs serve as a microcosm of real-world scenarios, equipping you with the skills to efficiently process and analyze text data. Time permitting, you?ll also explore modern tools like Python libraries, the OpenAI GPT-3 API, and TensorFlow, using them in a series of engaging exercises. By the end of the course, you'll have a well-rounded understanding of NLP, and will leave equipped with the practical skills and insights that you can immediately put to use, helping your organization gain valuable insights from text data, streamline business processes, and improve user interactions with automated text-based systems. You?ll be able to process and analyze text data effectively, implement advanced text representations, apply machine learning algorithms for text data, and build simple chatbots. Launch into the Universe of Natural Language Processing The journey begins: Unravel the layers of NLP Navigating through the history of NLP Merging paths: Text Analytics and NLP Decoding language: Word Sense Disambiguation and Sentence Boundary Detection First steps towards an NLP Project Unleashing the Power of Feature Extraction Dive into the vast ocean of Data Types Purification process: Cleaning Text Data Excavating knowledge: Extracting features from Texts Drawing connections: Finding Text Similarity through Feature Extraction Engineer Your Text Classifier The new era of Machine Learning and Supervised Learning Architecting a Text Classifier Constructing efficient workflows: Building Pipelines for NLP Projects Ensuring continuity: Saving and Loading Models Master the Art of Web Scraping and API Usage Stepping into the digital world: Introduction to Web Scraping and APIs The great heist: Collecting Data by Scraping Web Pages Navigating through the maze of Semi-Structured Data Unearth Hidden Themes with Topic Modeling Embark on the path of Topic Discovery Decoding algorithms: Understanding Topic-Modeling Algorithms Dialing the right numbers: Key Input Parameters for LSA Topic Modeling Tackling complexity with Hierarchical Dirichlet Process (HDP) Delving Deep into Vector Representations The Geometry of Language: Introduction to Vectors in NLP Text Manipulation: Generation and Summarization Playing the creator: Generating Text with Markov Chains Distilling knowledge: Understanding Text Summarization and Key Input Parameters for TextRank Peering into the future: Recent Developments in Text Generation and Summarization Solving real-world problems: Addressing Challenges in Extractive Summarization Riding the Wave of Sentiment Analysis Unveiling emotions: Introduction to Sentiment Analysis Tools Demystifying the Textblob library Preparing the canvas: Understanding Data for Sentiment Analysis Training your own emotion detectors: Building Sentiment Models Optional: Capstone Project Apply the skills learned throughout the course. Define the problem and gather the data. Conduct exploratory data analysis for text data. Carry out preprocessing and feature extraction. Select and train a model. ? Evaluate the model and interpret the results. Bonus Chapter: Generative AI and NLP Introduction to Generative AI and its role in NLP. Overview of Generative Pretrained Transformer (GPT) models. Using GPT models for text generation and completion. Applying GPT models for improving autocomplete features. Use cases of GPT in question answering systems and chatbots. Bonus Chapter: Advanced Applications of NLP with GPT Fine-tuning GPT models for specific NLP tasks. Using GPT for sentiment analysis and text classification. Role of GPT in Named Entity Recognition (NER). Application of GPT in developing advanced chatbots. Ethics and limitations of GPT and generative AI technologies.

NLP Boot Camp / Hands-On Natural Language Processing  (TTAI3030)
Delivered OnlineFlexible Dates
Price on Enquiry

CertNexus Certified Data Science Practitioner (CDSP)

By Nexus Human

Duration 5 Days 30 CPD hours This course is intended for This course is designed for business professionals who leverage data to address business issues. The typical student in this course will have several years of experience with computing technology, including some aptitude in computer programming. However, there is not necessarily a single organizational role that this course targets. A prospective student might be a programmer looking to expand their knowledge of how to guide business decisions by collecting, wrangling, analyzing, and manipulating data through code; or a data analyst with a background in applied math and statistics who wants to take their skills to the next level; or any number of other data-driven situations. Ultimately, the target student is someone who wants to learn how to more effectively extract insights from their work and leverage that insight in addressing business issues, thereby bringing greater value to the business. Overview In this course, you will learn to: Use data science principles to address business issues. Apply the extract, transform, and load (ETL) process to prepare datasets. Use multiple techniques to analyze data and extract valuable insights. Design a machine learning approach to address business issues. Train, tune, and evaluate classification models. Train, tune, and evaluate regression and forecasting models. Train, tune, and evaluate clustering models. Finalize a data science project by presenting models to an audience, putting models into production, and monitoring model performance. For a business to thrive in our data-driven world, it must treat data as one of its most important assets. Data is crucial for understanding where the business is and where it's headed. Not only can data reveal insights, it can also inform?by guiding decisions and influencing day-to-day operations. This calls for a robust workforce of professionals who can analyze, understand, manipulate, and present data within an effective and repeatable process framework. In other words, the business world needs data science practitioners. This course will enable you to bring value to the business by putting data science concepts into practice Addressing Business Issues with Data Science Topic A: Initiate a Data Science Project Topic B: Formulate a Data Science Problem Extracting, Transforming, and Loading Data Topic A: Extract Data Topic B: Transform Data Topic C: Load Data Analyzing Data Topic A: Examine Data Topic B: Explore the Underlying Distribution of Data Topic C: Use Visualizations to Analyze Data Topic D: Preprocess Data Designing a Machine Learning Approach Topic A: Identify Machine Learning Concepts Topic B: Test a Hypothesis Developing Classification Models Topic A: Train and Tune Classification Models Topic B: Evaluate Classification Models Developing Regression Models Topic A: Train and Tune Regression Models Topic B: Evaluate Regression Models Developing Clustering Models Topic A: Train and Tune Clustering Models Topic B: Evaluate Clustering Models Finalizing a Data Science Project Topic A: Communicate Results to Stakeholders Topic B: Demonstrate Models in a Web App Topic C: Implement and Test Production Pipelines

CertNexus Certified Data Science Practitioner (CDSP)
Delivered OnlineFlexible Dates
Price on Enquiry

Big Data Architecture Workshop

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for Senior Executives CIOs and CTOs Business Intelligence Executives Marketing Executives Data & Business Analytics Specialists Innovation Specialists & Entrepreneurs Academics, and other people interested in Big Data Overview More specifically, BDAW addresses advanced big data architecture topics, including, data formats, transformation, real-time, batch and machine learning processing, scalability, fault tolerance, security and privacy, minimizing the risk of an unsound architecture and technology selection. Big Data Architecture Workshop (BDAW) is a learning event that addresses advanced big data architecture topics. BDAW brings together technical contributors into a group setting to design and architect solutions to a challenging business problem. The workshop addresses big data architecture problems in general, and then applies them to the design of a challenging system. Throughout the highly interactive workshop, students apply concepts to real-world examples resulting in detailed synergistic discussions. The workshop is conducive for students to learn techniques for architecting big data systems, not only from Cloudera?s experience but also from the experiences of fellow students. Workshop Application Use Cases Oz Metropolitan Architectural questions Team activity: Analyze Metroz Application Use Cases Application Vertical Slice Definition Minimizing risk of an unsound architecture Selecting a vertical slice Team activity: Identify an initial vertical slice for Metroz Application Processing Real time, near real time processing Batch processing Data access patterns Delivery and processing guarantees Machine Learning pipelines Team activity: identify delivery and processing patterns in Metroz, characterize response time requirements, identify Machine Learning pipelines Application Data Three V?s of Big Data Data Lifecycle Data Formats Transforming Data Team activity: Metroz Data Requirements Scalable Applications Scale up, scale out, scale to X Determining if an application will scale Poll: scalable airport terminal designs Hadoop and Spark Scalability Team activity: Scaling Metroz Fault Tolerant Distributed Systems Principles Transparency Hardware vs. Software redundancy Tolerating disasters Stateless functional fault tolerance Stateful fault tolerance Replication and group consistency Fault tolerance in Spark and Map Reduce Application tolerance for failures Team activity: Identify Metroz component failures and requirements Security and Privacy Principles Privacy Threats Technologies Team activity: identify threats and security mechanisms in Metroz Deployment Cluster sizing and evolution On-premise vs. Cloud Edge computing Team activity: select deployment for Metroz Technology Selection HDFS HBase Kudu Relational Database Management Systems Map Reduce Spark, including streaming, SparkSQL and SparkML Hive Impala Cloudera Search Data Sets and Formats Team activity: technologies relevant to Metroz Software Architecture Architecture artifacts One platform or multiple, lambda architecture Team activity: produce high level architecture, selected technologies, revisit vertical slice Vertical Slice demonstration Additional course details: Nexus Humans Big Data Architecture Workshop training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Big Data Architecture Workshop course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Big Data Architecture Workshop
Delivered OnlineFlexible Dates
Price on Enquiry

Building Recommendation Systems with Python (TTAI2360)

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This course is geared for Python experienced developers, analysts or others who are intending to learn the tools and techniques required in building various kinds of powerful recommendation systems (collaborative, knowledge and content based) and deploying them to the web. Overview This skills-focused combines engaging lecture, demos, group activities and discussions with machine-based student labs and exercises.. Our engaging instructors and mentors are highly-experienced practitioners who bring years of current, modern 'on-the-job' modern applied datascience, AI and machine learning experience into every classroom and hands-on project. Working in a hands-on lab environment led by our expert instructor, attendees will Understand the different kinds of recommender systems Master data-wrangling techniques using the pandas library Building an IMDB Top 250 Clone Build a content-based engine to recommend movies based on real movie metadata Employ data-mining techniques used in building recommenders Build industry-standard collaborative filters using powerful algorithms Building Hybrid Recommenders that incorporate content based and collaborative filtering Recommendation systems are at the heart of almost every internet business today; from Facebook to Net?ix to Amazon. Providing good recommendations, whether its friends, movies, or groceries, goes a long way in defining user experience and enticing your customers to use your platform.This course shows you how to do just that. You will learn about the different kinds of recommenders used in the industry and see how to build them from scratch using Python. No need to wade through tons of machine learning theory?you will get started with building and learning about recommenders as quickly as possible. In this course, you will build an IMDB Top 250 clone, a content-based engine that works on movie metadata. You will also use collaborative filters to make use of customer behavior data, and a Hybrid Recommender that incorporates content based and collaborative filtering techniques. Students will learn to build industry-standard recommender systems, leveraging basic Python syntax skills. This is an applied course, so machine learning theory is only used to highlight how to build recommenders in this course. Getting Started with Recommender Systems Technical requirements What is a recommender system? Types of recommender systems Manipulating Data with the Pandas Library Technical requirements Setting up the environment The Pandas library The Pandas DataFrame The Pandas Series Building an IMDB Top 250 Clone with Pandas Technical requirements The simple recommender The knowledge-based recommender Building Content-Based Recommenders Technical requirements Exporting the clean DataFrame Document vectors The cosine similarity score Plot description-based recommender Metadata-based recommender Suggestions for improvements Getting Started with Data Mining Techniques Problem statement Similarity measures Clustering Dimensionality reduction Supervised learning Evaluation metrics Building Collaborative Filters Technical requirements The framework User-based collaborative filtering Item-based collaborative filtering Model-based approaches Hybrid Recommenders Technical requirements Introduction Case study and final project ? Building a hybrid model Additional course details: Nexus Humans Building Recommendation Systems with Python (TTAI2360) training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the Building Recommendation Systems with Python (TTAI2360) course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

Building Recommendation Systems with Python (TTAI2360)
Delivered OnlineFlexible Dates
Price on Enquiry

Google Cloud Fundamentals for Azure Professionals

By Nexus Human

Duration 1 Days 6 CPD hours This course is intended for This course is intended for the following participants: Individuals planning to deploy applications and create application environments on Google Cloud Platform Developers, systems operations professionals, and solution architects getting started with Google Cloud Platform Executives and business decision makers evaluating the potential of Google Cloud Platform to address their business needs. Overview This course teaches participants the following skills: Identify Google Cloud counterparts for Azure IaaS, Azure PaaS, Azure SQL, Azure Blob Storage, Azure Application Insights, and Azure Data Lake Configure accounts, billing, projects, networks, subnets, firewalls, VMs, disks, auto scaling, load balancing,storage, databases, IAM, and more Manage and monitor applications Explain feature and pricing model differences This 1-day instructor led course introduces Azure professionals to the core capabilities of Google Cloud in the four technology pillars: networking, compute, storage, and database. It is designed for Azure system administrators, Solution Architects and SysOps Administrators familiar with Azure features and setup; and want to gain experience configuring Google Cloud products immediately. With presentations, demos, and hands-on labs, participants get details of similarities, differences, and initial how-tos quickly. Introducing Google Cloud Explain the advantages of Google Cloud. Define the components of Google's network infrastructure, including: Points of presence, data centers, regions, and zones. Understand the difference between Infrastructure-as-a-Service (IaaS) and Platform-as-a-Service (PaaS). Getting Started with Google Cloud Identify the purpose of projects on Google Cloud. Understand how Azure's resource hierarchy differs from Google Cloud's Understand the purpose of and use cases for Identity and Access Management. Understand how Azure AD differs from Google Cloud IAM. List the methods of interacting with Google Cloud. Launch a solution using Cloud Marketplace. Virtual Machines in the Cloud Identify the purpose and use cases for Google Compute Engine Understand the basics of networking in Google Cloud. Understand how Azure VPC differs from Google VPC. Understand the similarities and differences between Azure VM and Google Compute Engine. Understand how typical approaches to load-balancing in Google Cloud differ from those in Azure. Deploy applications using Google Compute Engine Storage in the Cloud Understand the purpose of and use cases for: Cloud Storage, Cloud SQL, Cloud Bigtable and Cloud Datastore. Understand how Azure Blob compares to Cloud Storage. Compare Google Cloud?s managed database services with Azure SQL. Learn how to choose among the various storage options on Google Cloud. Load data from Cloud Storage into BigQuery Containers in the Cloud Define the concept of a container and identify uses for containers. Identify the purpose of and use cases for Google Container Engine and Kubernetes. Understand how Azure Kubernetes Service differs from from Google Kubernetes Engine. Provision a Kubernetes cluster using Kubernetes Engine. Deploy and manage Docker containers using kubectl Applications in the Cloud Understand the purpose of and use cases for Google App Engine. Contrast the App Engine Standard environment with the App Engine Flexible environment. Understand how App Engine differs from Azure App Service. Understand the purpose of and use cases for Google Cloud Endpoints. Developing, Deploying and Monitoring in the Cloud Understand options for software developers to host their source code. Understand the purpose of template-based creation and management of resources. Understand how Google Cloud Deployment Manager differs from Azure Resource Manager. Understand the purpose of integrated monitoring, alerting, and debugging Understand how Google Monitoring differs from Azure Application Insights and Azure Log Analytics. Create a Deployment Manager deployment. Update a Deployment Manager deployment. View the load on a VM instance using Google Monitoring. Big Data and Machine Learning in the Cloud Understand the purpose of and use cases for the products and services in the Google Cloud big data and machine learning platforms. Understand how Google Cloud BigQuery differs from Azure Data Lake. Understand how Google Cloud Pub/Sub differs from Azure Event Hubs and Service Bus. Understand how Google Cloud?s machine-learning APIs differ from Azure's. Load data into BigQuery from Cloud Storage. Perform queries using BigQuery to gain insight into data Summary and Review Review the products that make up Google Cloud and remember how to choose among them Understand next steps for training and certification Understand, at a high level, the process of migrating from Azure to Google Cloud.

Google Cloud Fundamentals for Azure Professionals
Delivered OnlineFlexible Dates
Price on Enquiry

Fast Track to Scala Programming for OO / Java Developers (TTSCL2104)

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This course is geared for experienced skilled Java developers, software developers, data scientists, machine learning experts or others who wish to transtion their coding skills to Scala, learning how to code in Scala and apply it in a practical way. This is not a basic class. Overview Working in a hands-on learning environment led by our expert instructor you'll: Get comfortable with Scala's core principles and unique features, helping you navigate the language confidently and boosting your programming skills. Discover the power of functional programming and learn techniques that will make your code more efficient,maintainable, and enjoyable to write. Become proficient in creating dynamic web applications using the Play Framework, and easily connect to databases with the user-friendly Slick library. Master concurrency programming with Akka, empowering you to build scalable and fault-tolerant applications that excel in performance. Enhance your testing skills using ScalaTest and ScalaCheck, ensuring the reliability and quality of your Scala applications, while having fun in the process. Explore the fascinating world of generative AI and GPT technologies, and learn how to integrate them into your projects,adding a touch of innovation and intelligence to your Scala solutions. If your team requires different topics, additional skills or a custom approach, our team will collaborate with you to adjust the course to focus on your specific learning objectives and goals. Discover the power of Scala programming in our comprehensive, hands-on technical training course designed specifically for experienced object-oriented (OO) developers. Scala is a versatile programming language that combines the best of both OO and functional programming paradigms, making it ideal for a wide range of projects, from web applications to big data processing and machine learning. By mastering Scala, you'll be able to develop more efficient, scalable, and maintainable applications. Fast Track to Scala Programming for OO / Java Developers is a four day hands-on course covers the core principles of Scala, functional programming, web application development, database connectivity, concurrency programming, testing, and interoperability between Scala and Java. Additionally, you'll explore cutting-edge generative AI and GPT technologies, learning how to integrate them into your Scala applications for intelligent suggestions or automation. Throughout the course you?ll explore the latest tools and best practices in the Scala ecosystem, gaining valuable knowledge and experience that can be directly applied to your day-to-day work. With 50% of the course content dedicated to hands-on labs, you'll gain practical experience applying the concepts you've learned across various projects, such as building functional web applications, connecting to databases, designing modular components, and implementing concurrency. Upon completing the course, you'll have a solid understanding of the language and its features, empowering you to confidently apply your new skills in data science and machine learning projects. You'll exit well-prepared to create efficient, scalable, and maintainable Scala applications, regardless of the complexity of your projects. Introduction to Scala Scala features and benefits Comparing Scala with Java and other OO languages Installing Scala and setting up the development environment Object-Oriented Programming in Scala Classes and objects Traits, mixins, and inheritance Companion objects and factories Encapsulation and polymorphism Functional Programming Basics Pure functions and referential transparency Higher-order functions and currying Immutability and persistent data structures Pattern matching and recursion Having Fun with Functional Data Structures Lists, sets, and maps in Scala Folding and reducing operations Stream processing and lazy evaluation For-comprehensions Building Web Applications in Functional Style Introduction to Play Framework Functional web routing and request handling JSON handling with Play-JSON Middleware and functional composition Connecting to a Database Introduction to Slick library Database configuration and setup Querying and updating with Slick Transactions and error handling Building Scalable and Extensible Components Modular architecture and design patterns Dependency injection with MacWire Type classes and type-level programming Implicit parameters and conversions Concurrency Programming & Akka Introduction to Akka framework and Actor model Actor systems and message passing Futures and Promises Supervision and fault tolerance Building Confidence with Testing Introduction to ScalaTest and ScalaCheck Unit testing and property-based testing Test-driven development in Scala Mocking and integration testing Interoperability between Scala and Java Calling Java code from Scala Using Java libraries in Scala projects Converting Java collections to Scala collections Writing Scala code that can be called from Java Using Generative AI and GPT Technologies in Scala Programming Overview of GPT and generative AI Integrating GPT with Scala applications Use cases and practical examples

Fast Track to Scala Programming for OO / Java Developers (TTSCL2104)
Delivered OnlineFlexible Dates
Price on Enquiry

Data Engineering on Google Cloud

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for This class is intended for experienced developers who are responsible for managing big data transformations including: Extracting, loading, transforming, cleaning, and validating data. Designing pipelines and architectures for data processing. Creating and maintaining machine learning and statistical models. Querying datasets, visualizing query results and creating reports Overview Design and build data processing systems on Google Cloud Platform. Leverage unstructured data using Spark and ML APIs on Cloud Dataproc. Process batch and streaming data by implementing autoscaling data pipelines on Cloud Dataflow. Derive business insights from extremely large datasets using Google BigQuery. Train, evaluate and predict using machine learning models using TensorFlow and Cloud ML. Enable instant insights from streaming data Get hands-on experience with designing and building data processing systems on Google Cloud. This course uses lectures, demos, and hand-on labs to show you how to design data processing systems, build end-to-end data pipelines, analyze data, and implement machine learning. This course covers structured, unstructured, and streaming data. Introduction to Data Engineering Explore the role of a data engineer. Analyze data engineering challenges. Intro to BigQuery. Data Lakes and Data Warehouses. Demo: Federated Queries with BigQuery. Transactional Databases vs Data Warehouses. Website Demo: Finding PII in your dataset with DLP API. Partner effectively with other data teams. Manage data access and governance. Build production-ready pipelines. Review GCP customer case study. Lab: Analyzing Data with BigQuery. Building a Data Lake Introduction to Data Lakes. Data Storage and ETL options on GCP. Building a Data Lake using Cloud Storage. Optional Demo: Optimizing cost with Google Cloud Storage classes and Cloud Functions. Securing Cloud Storage. Storing All Sorts of Data Types. Video Demo: Running federated queries on Parquet and ORC files in BigQuery. Cloud SQL as a relational Data Lake. Lab: Loading Taxi Data into Cloud SQL. Building a Data Warehouse The modern data warehouse. Intro to BigQuery. Demo: Query TB+ of data in seconds. Getting Started. Loading Data. Video Demo: Querying Cloud SQL from BigQuery. Lab: Loading Data into BigQuery. Exploring Schemas. Demo: Exploring BigQuery Public Datasets with SQL using INFORMATION_SCHEMA. Schema Design. Nested and Repeated Fields. Demo: Nested and repeated fields in BigQuery. Lab: Working with JSON and Array data in BigQuery. Optimizing with Partitioning and Clustering. Demo: Partitioned and Clustered Tables in BigQuery. Preview: Transforming Batch and Streaming Data. Introduction to Building Batch Data Pipelines EL, ELT, ETL. Quality considerations. How to carry out operations in BigQuery. Demo: ELT to improve data quality in BigQuery. Shortcomings. ETL to solve data quality issues. Executing Spark on Cloud Dataproc The Hadoop ecosystem. Running Hadoop on Cloud Dataproc. GCS instead of HDFS. Optimizing Dataproc. Lab: Running Apache Spark jobs on Cloud Dataproc. Serverless Data Processing with Cloud Dataflow Cloud Dataflow. Why customers value Dataflow. Dataflow Pipelines. Lab: A Simple Dataflow Pipeline (Python/Java). Lab: MapReduce in Dataflow (Python/Java). Lab: Side Inputs (Python/Java). Dataflow Templates. Dataflow SQL. Manage Data Pipelines with Cloud Data Fusion and Cloud Composer Building Batch Data Pipelines visually with Cloud Data Fusion. Components. UI Overview. Building a Pipeline. Exploring Data using Wrangler. Lab: Building and executing a pipeline graph in Cloud Data Fusion. Orchestrating work between GCP services with Cloud Composer. Apache Airflow Environment. DAGs and Operators. Workflow Scheduling. Optional Long Demo: Event-triggered Loading of data with Cloud Composer, Cloud Functions, Cloud Storage, and BigQuery. Monitoring and Logging. Lab: An Introduction to Cloud Composer. Introduction to Processing Streaming Data Processing Streaming Data. Serverless Messaging with Cloud Pub/Sub Cloud Pub/Sub. Lab: Publish Streaming Data into Pub/Sub. Cloud Dataflow Streaming Features Cloud Dataflow Streaming Features. Lab: Streaming Data Pipelines. High-Throughput BigQuery and Bigtable Streaming Features BigQuery Streaming Features. Lab: Streaming Analytics and Dashboards. Cloud Bigtable. Lab: Streaming Data Pipelines into Bigtable. Advanced BigQuery Functionality and Performance Analytic Window Functions. Using With Clauses. GIS Functions. Demo: Mapping Fastest Growing Zip Codes with BigQuery GeoViz. Performance Considerations. Lab: Optimizing your BigQuery Queries for Performance. Optional Lab: Creating Date-Partitioned Tables in BigQuery. Introduction to Analytics and AI What is AI?. From Ad-hoc Data Analysis to Data Driven Decisions. Options for ML models on GCP. Prebuilt ML model APIs for Unstructured Data Unstructured Data is Hard. ML APIs for Enriching Data. Lab: Using the Natural Language API to Classify Unstructured Text. Big Data Analytics with Cloud AI Platform Notebooks What's a Notebook. BigQuery Magic and Ties to Pandas. Lab: BigQuery in Jupyter Labs on AI Platform. Production ML Pipelines with Kubeflow Ways to do ML on GCP. Kubeflow. AI Hub. Lab: Running AI models on Kubeflow. Custom Model building with SQL in BigQuery ML BigQuery ML for Quick Model Building. Demo: Train a model with BigQuery ML to predict NYC taxi fares. Supported Models. Lab Option 1: Predict Bike Trip Duration with a Regression Model in BQML. Lab Option 2: Movie Recommendations in BigQuery ML. Custom Model building with Cloud AutoML Why Auto ML? Auto ML Vision. Auto ML NLP. Auto ML Tables.

Data Engineering on Google Cloud
Delivered OnlineFlexible Dates
Price on Enquiry