Big Data
Showing 25–36 of 83 results
Data Science in Production: Building Scalable Model Pipelines
Learn how to build scalable model pipelines for data science in production environments, focusing on best practices for deployment and management.
DelftX: Modern Distributed Systems
Distributed systems are the backbone of modern society but entail challenges in areas such as complexity and energy-use. Discover distributed systems from first principles, understand the architectures and techniques derived from them and explore examples of current practical use.
DelftX: Unix Tools: Data, Software and Production Engineering
Grow from being a Unix novice to Unix wizard status! Process big data, analyze software code, run DevOps tasks and excel in your everyday job through the amazing power of the Unix shell and command-line tools.
Deploying a Hadoop Cluster
Gain practical knowledge on how to deploy and manage a Hadoop cluster for big data processing.
Developing Distributed Applications Using ZooKeeper
Understand how to develop reliable distributed applications with Apache ZooKeeper. Learn how to manage configurations, synchronization, and group services effectively.
EdinburghX: Data Ethics, AI and Responsible Innovation
Our future is here and it relies on data. Medical robots, smart homes and cities, predictive policing, artificial intelligences – all are fuelled by data and all promise new benefits to society. But will these innovations benefit everyone? Who stands to gain and who is put at risk? How can we ensure that data is part of a just and sustainable world?
Enterprise Skills in Hortonworks Data Platform
Hortonworks is the top Big Data Platforms that is 100% opensource built for the Big Data community. This course is going to focus on the Enterprise Skills needed for Data Engineers to build and support critical application in the Hadoop Ecosystem.
GeorgetownX: Demystifying Biomedical Big Data: A User’s Guide
Whether you are a student, basic scientist, researcher, clinician, or librarian, this course is designed to help you understand, analyze, and interpret biomedical big data.
Getting Started with Apache Kafka
Discover the immense value of Kafka as a messaging platform, trusted by global giants. Gain insights on configuring it for high throughputs without sacrificing ordering guarantee.
Getting Started with Enterprise Search Using Apache Solr
Search is one of the most misunderstood functionalities in the IT industry. Apache Solr brings high quality Enterprise Search to the masses.
Getting Started with HBase: The Hadoop Database
As the data you store expands in size, traditional relational databases may no longer work. HBase has the ability to deal with billions of rows of data and each record can contains millions of fields. This course will help you get started with HBase.
Getting Started with HDFS
Learning to work with Hadoop Distributed File System (HDFS) is a baseline skill for anyone administering or developing in the Hadoop ecosystem. In this course, you will learn how to work with HDFS, Hive, Pig, Sqoop and HBase from the command line.