We couldn't find any listings for your search. Explore our online options below.
Know someone teaching this? Help them become an Educator on Cademy.
Apache NiFi, a robust, open-source data ingestion/distribution framework, is the core of Hortonworks DataFlow (HDF)
This course will help you explore the world of Big Data technologies and frameworks. You will develop skills that will help you to pick the right Big Data technology and framework for your job and build the confidence to design robust Big Data pipelines.
Get hands-on with Kafka monitoring setup with Prometheus and Grafana, Kafka operations and Kafka cluster upgrades Setup in AWS.
Better than REST APIs! Build a fast and scalable HTTP/2 API for a Go microservice with gRPC and protocol buffers (protobufs)
Better than REST APIs! Build a fast and scalable HTTP/2 API for your microservice with gRPC and protocol buffers (protobufs).
With this course, you will master all CloudFormation concepts, and become confident in writing CloudFormation templates using YAML. Throughout the course, you will encounter various interesting examples and activities that will help you to consolidate your learning.
A beginner-level course that follows a step-by-step approach to learning the fundamentals and core concepts of Apache Kafka 3.0. You will work through interesting activities such as programming a Twitter producer and Elasticsearch consumer to understand the various concepts.
This course covers the important topics needed to pass the AWS Certified Data Analytics-Specialty exam (AWS DAS-C01). You will learn about Kinesis, EMR, DynamoDB, and Redshift, and get ready for the exam by working through quizzes, exercises, and practice exams, along with exploring essential tips and techniques.
This course will show you why Hadoop is one of the best tools to work with big data. With the help of some real-world data sets, you will learn how to use Hadoop and its distributed technologies, such as Spark, Flink, Pig, and Flume, to store, analyze, and scale big data.