• Professional Development
  • Medicine & Nursing
  • Arts & Crafts
  • Health & Wellbeing
  • Personal Development

5 Educators providing ODI courses in Bolsover delivered Live Online

Courses matching "ODI"

Show all 4

Oracle Data Integrator 19c Configuration and Administration (TTOR30319)

By Nexus Human

Duration 3 Days 18 CPD hours This course is intended for This intermediate-level hands-on course is geared for experienced Administrators, Analysts, Architects, Data Scientists, Database Administrators and Implementers Overview This course is approximately 50% hands-on, combining expert lecture, real-world demonstrations and group discussions with machine-based practical labs and exercises. Working in a hands-on learning environment led by our Oracle Certified expert facilitator, students will learn how to: Administer ODI resources and setup security with ODI Apply ODI Topology concepts for data integration Describe ODI Model concepts Describe architecture of Oracle Data Integrator Design ODI Mappings, Procedures, Packages, and Load Plans to perform ELT data transformations Explore, audit data, and enforce data quality with ODI Implement Changed Data Capture with ODI Oracle Data Integrator is a comprehensive data integration platform that covers all data integration requirements from high-volume, high-performance batch loads, to event-driven integration processes and SOA-enabled data services. Oracle Data Integrator's Extract, Load, Transform (E-LT) architecture leverages disparate RDBMS engines to process and transform the data - the approach that optimizes performance, scalability and lowers overall solution costs. Throughout this course participants will explore how to centralize data across databases, performing integration, designing ODI Mappings, and setting up ODI security. In addition, Oracle Data Integrator can interact with the various tools of the Hadoop ecosystem, allowing administrators and data scientists to farm out map-reduce operations from established relational databases to Hadoop. They can also read back into the relational world the results of complex Big Data analysis for further processing. Working in a hands-on learning environment led by our Oracle Certified expert facilitator, students will learn how to: Administer ODI resources and setup security with ODI Apply ODI Topology concepts for data integration Describe ODI Model concepts Describe architecture of Oracle Data Integrator Design ODI Mappings, Procedures, Packages, and Load Plans to perform ELT data transformations Explore, audit data, and enforce data quality with ODI Implement Changed Data Capture with ODI Introduction to Integration and Administration Oracle Data Integrator: Introduction Oracle Data Integrator Repositories Administering ODI Repositories Create and connect to the Master Repository Export and import the Master Repository Create, connect, and set a password to the Work Repository ODI Topology Concepts ODI Topology: Overview Data Servers and Physical Schemas Defining Topology Agents in Topology Planning a Topology Describing the Physical and Logical Architecture Topology Navigator Creating Physical Architecture Creating Logical Architecture Setting Up a New ODI Project ODI Projects Using Folders Understanding Knowledge Modules Exporting and Importing Objects Using Markers Oracle Data Integrator Model Concepts Understanding the Relational Model Understanding Reverse-Engineering Creating Models Organizing ODI Models and Creating ODI Datastores Organizing Models Creating Datastores Constraints in ODI Creating Keys and References Creating Conditions Exploring Your Data Constructing Business Rules ODI Mapping Concepts ODI Mappings Expressions, Join, Filter, Lookup, Sets, and Others Behind the Rules Staging Area and Execution Location Understanding Knowledge Modules Mappings: Overview Designing Mappings Multiple Sources and Joins Filtering Data Overview of the Flow in ODI Mapping Selecting a Staging Area Configuring Expressions Execution Location Selecting a Knowledge Module Mappings: Monitoring and Troubleshooting Monitoring Mappings Working with Errors Designing Mappings: Advanced Topics 1 Working with Business Rules Using Variables Datasets and Sets Using Sequences Designing Mappings: Advanced Topics 2 Partitioning Configuring Reusable Mappings Using User Functions Substitution Methods Modifying Knowledge Modules Using ODI Procedures Procedures: Overview Creating a Blank Procedure Adding Commands Adding Options Running a Procedure Using ODI Packages Packages: Overview Executing a Package Review of Package Steps Model, Submodel, and Datastore Steps Variable Steps Controlling the Execution Path Step-by-Step Debugger Starting a Debug Session New Functions Menu Bar Icons Managing ODI Scenarios Scenarios Managing Scenarios Preparing for Deployment Using Load Plans What are load plans? Load plan editor Load plan step sequence Defining restart behavior Enforcing Data Quality with ODI Data Quality Business Rules for Data Quality Enforcing Data Quality with ODI Working with Changed Data Capture CDC with ODI CDC implementations with ODI CDC implementation techniques Journalizing Results of CDC Advanced ODI Administration Setting Up ODI Security Managing ODI Reports ODI Integration with Java

Oracle Data Integrator 19c Configuration and Administration (TTOR30319)
Delivered OnlineFlexible Dates
Price on Enquiry

55321 SQL Server Integration Services

By Nexus Human

Duration 5 Days 30 CPD hours This course is intended for The primary audience for this course is database professionals who need to fulfil a Business Intelligence Developer role. They will need to focus on hands-on work creating BI solutions including Data Warehouse implementation, ETL, and data cleansing. Overview Create sophisticated SSIS packages for extracting, transforming, and loading data Use containers to efficiently control repetitive tasks and transactions Configure packages to dynamically adapt to environment changes Use Data Quality Services to cleanse data Successfully troubleshoot packages Create and Manage the SSIS Catalog Deploy, configure, and schedule packages Secure the SSIS Catalog SQL Server Integration Services is the Community Courseware version of 20767CC Implementing a SQL Data Warehouse. This five-day instructor-led course is intended for IT professionals who need to learn how to use SSIS to build, deploy, maintain, and secure Integration Services projects and packages, and to use SSIS to extract, transform, and load data to and from SQL Server. This course is similar to the retired Course 20767-C: Implementing a SQL Data Warehouse but focuses more on building packages, rather than the entire data warehouse design and implementation. Prerequisites Working knowledge of T-SQL and SQL Server Agent jobs is helpful, but not required. Basic knowledge of the Microsoft Windows operating system and its core functionality. Working knowledge of relational databases. Some experience with database design. 1 - SSIS Overview Import/Export Wizard Exporting Data with the Wizard Common Import Concerns Quality Checking Imported/Exported Data 2 - Working with Solutions and Projects Working with SQL Server Data Tools Understanding Solutions and Projects Working with the Visual Studio Interface 3 - Basic Control Flow Working with Tasks Understanding Precedence Constraints Annotating Packages Grouping Tasks Package and Task Properties Connection Managers Favorite Tasks 4 - Common Tasks Analysis Services Processing Data Profiling Task Execute Package Task Execute Process Task Expression Task File System Task FTP Task Hadoop Task Script Task Introduction Send Mail Task Web Service Task XML Task 5 - Data Flow Sources and Destinations The Data Flow Task The Data Flow SSIS Toolbox Working with Data Sources SSIS Data Sources Working with Data Destinations SSIS Data Destinations 6 - Data Flow Transformations Transformations Configuring Transformations 7 - Making Packages Dynamic Features for Making Packages Dynamic Package Parameters Project Parameters Variables SQL Parameters Expressions in Tasks Expressions in Connection Managers After Deployment How It All Fits Together 8 - Containers Sequence Containers For Loop Containers Foreach Loop Containers 9 - Troubleshooting and Package Reliability Understanding MaximumErrorCount Breakpoints Redirecting Error Rows Logging Event Handlers Using Checkpoints Transactions 10 - Deploying to the SSIS Catalog The SSIS Catalog Deploying Projects Working with Environments Executing Packages in SSMS Executing Packages from the Command Line Deployment Model Differences 11 - Installing and Administering SSIS Installing SSIS Upgrading SSIS Managing the SSIS Catalog Viewing Built-in SSIS Reports Managing SSIS Logging and Operation Histories Automating Package Execution 12 - Securing the SSIS Catalog Principals Securables Grantable Permissions Granting Permissions Configuring Proxy Accounts Additional course details: Nexus Humans 55321 SQL Server Integration Services training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the 55321 SQL Server Integration Services course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

55321 SQL Server Integration Services
Delivered OnlineFlexible Dates
£2,975

KM204 IBM InfoSphere DataStage Essentials (v11.5)

By Nexus Human

Duration 4 Days 24 CPD hours This course is intended for Project administrators and ETL developers responsible for data extraction and transformation using DataStage. Overview Describe the uses of DataStage and the DataStage workflowDescribe the Information Server architecture and how DataStage fits within itDescribe the Information Server and DataStage deployment optionsUse the Information Server Web Console and the DataStage Administrator client to create DataStage users and to configure the DataStage environmentImport and export DataStage objects to a fileImport table definitions for sequential files and relational tablesDesign, compile, run, and monitor DataStage parallel jobsDesign jobs that read and write to sequential filesDescribe the DataStage parallel processing architectureDesign jobs that combine data using joins and lookupsDesign jobs that sort and aggregate dataImplement complex business logic using the DataStage Transformer stageDebug DataStage jobs using the DataStage PX Debugger This course enables the project administrators & developers to acquire the skills necessary to develop parallel jobs in DataStage. Students will learn to create parallel jobs that access sequential & relational data, and combine and transform the data. Course Outline Introduction to DataStage Deployment DataStage Administration Work with Metadata Create Parallel Jobs Access Sequential Data Partitioning and Collecting Algorithms Combine Data Group Processing Stages Transformer Stage Repository Functions Work with Relational Data Control Jobs

KM204 IBM InfoSphere DataStage Essentials (v11.5)
Delivered OnlineFlexible Dates
Price on Enquiry

KM423 IBM InfoSphere DataStage v11.5 - Advanced Data Processing

By Nexus Human

Duration 2 Days 12 CPD hours This course is intended for Experienced DataStage developers seeking training in more advanced DataStage job techniques and who seek techniques for working with complex types of data resources. Overview Use Connector stages to read from and write to database tables Handle SQL errors in Connector stages Use Connector stages with multiple input links Use the File Connector stage to access Hadoop HDFS data Optimize jobs that write to database tables Use the Unstructured Data stage to extract data from Excel spreadsheets Use the Data Masking stage to mask sensitive data processed within a DataStage job Use the Hierarchical stage to parse, compose, and transform XML data Use the Schema Library Manager to import and manage XML schemas Use the Data Rules stage to validate fields of data within a DataStage job Create custom data rules for validating data Design a job that processes a star schema data warehouse with Type 1 and Type 2 slowly changing dimensions This course is designed to introduce you to advanced parallel job data processing techniques in DataStage v11.5. In this course you will develop data techniques for processing different types of complex data resources including relational data, unstructured data (Excel spreadsheets), and XML data. In addition, you will learn advanced techniques for processing data, including techniques for masking data and techniques for validating data using data rules. Finally, you will learn techniques for updating data in a star schema data warehouse using the DataStage SCD (Slowly Changing Dimensions) stage. Even if you are not working with all of these specific types of data, you will benefit from this course by learning advanced DataStage job design techniques, techniques that go beyond those utilized in the DataStage Essentials course. Accessing databases Connector stage overview - Use Connector stages to read from and write to relational tables - Working with the Connector stage properties Connector stage functionality - Before / After SQL - Sparse lookups - Optimize insert/update performance Error handling in Connector stages - Reject links - Reject conditions Multiple input links - Designing jobs using Connector stages with multiple input links - Ordering records across multiple input links File Connector stage - Read and write data to Hadoop file systems Demonstration 1: Handling database errors Demonstration 2: Parallel jobs with multiple Connector input links Demonstration 3: Using the File Connector stage to read and write HDFS files Processing unstructured data Using the Unstructured Data stage in DataStage jobs - Extract data from an Excel spreadsheet - Specify a data range for data extraction in an Unstructured Data stage - Specify document properties for data extraction. Demonstration 1: Processing unstructured data Data masking Using the Data Masking stage in DataStage jobs - Data masking techniques - Data masking policies - Applying policies for masquerading context-aware data types - Applying policies for masquerading generic data types - Repeatable replacement - Using reference tables - Creating custom reference tables Demonstration 1: Data masking Using data rules Introduction to data rules - Using the Data Rules Editor - Selecting data rules - Binding data rule variables - Output link constraints - Adding statistics and attributes to the output information Use the Data Rules stage to valid foreign key references in source data Create custom data rules Demonstration 1: Using data rules Processing XML data Introduction to the Hierarchical stage - Hierarchical stage Assembly editor - Use the Schema Library Manager to import and manage XML schemas Composing XML data - Using the HJoin step to create parent-child relationships between input lists - Using the Composer step Writing Hierarchical data to a relational table Using the Regroup step Consuming XML data - Using the XML Parser step - Propagating columns Topic 6: Transforming XML data - Using the Aggregate step - Using the Sort step - Using the Switch step - Using the H-Pivot step Demonstration 1: Importing XML schemas Demonstration 2: Compose hierarchical data Demonstration 3: Consume hierarchical data Demonstration 4: Transform hierarchical data Updating a star schema database Surrogate keys - Design a job that creates and updates a surrogate key source key file from a dimension table Slowly Changing Dimensions (SCD) stage - Star schema databases - SCD stage Fast Path pages - Specifying purpose codes - Dimension update specification - Design a job that processes a star schema database with Type 1 and Type 2 slowly changing dimensions Demonstration 1: Build a parallel job that updates a star schema database with two dimensions Additional course details: Nexus Humans KM423 IBM InfoSphere DataStage v11.5 - Advanced Data Processing training program is a workshop that presents an invigorating mix of sessions, lessons, and masterclasses meticulously crafted to propel your learning expedition forward. This immersive bootcamp-style experience boasts interactive lectures, hands-on labs, and collaborative hackathons, all strategically designed to fortify fundamental concepts. Guided by seasoned coaches, each session offers priceless insights and practical skills crucial for honing your expertise. Whether you're stepping into the realm of professional skills or a seasoned professional, this comprehensive course ensures you're equipped with the knowledge and prowess necessary for success. While we feel this is the best course for the KM423 IBM InfoSphere DataStage v11.5 - Advanced Data Processing course and one of our Top 10 we encourage you to read the course outline to make sure it is the right content for you. Additionally, private sessions, closed classes or dedicated events are available both live online and at our training centres in Dublin and London, as well as at your offices anywhere in the UK, Ireland or across EMEA.

KM423 IBM InfoSphere DataStage v11.5 - Advanced Data Processing
Delivered OnlineFlexible Dates
Price on Enquiry