Customize your search:

E.g., 2017-07-18
E.g., 2017-07-18
E.g., 2017-07-18
Jul 24th 2017

This course is for those new to data science.

Average: 6.8 (6 votes)
Jul 24th 2017

This course is for novice programmers or business people who'd like to understand the core tools used to wrangle and analyze big data. With no prior experience, you'll have the opportunity to walk through hands-on examples with Hadoop and Spark frameworks, two of the most common in the industry. You will be comfortable explaining the specific components and basic processes of the Hadoop architecture, software stack, and execution environment. In the assignments you will be guided in how data scientists apply the important concepts and techniques, such as Map-Reduce that are used to solve fundamental problems in big data. You'll feel empowered to have conversations about big data and the data analysis processes.

Average: 6.6 (10 votes)
Jul 17th 2017

Welcome to the Cloud Computing Applications course, the second part of a two-course series designed to give you a comprehensive view on the world of Cloud Computing and Big Data!

Average: 7.4 (5 votes)

Jul 17th 2017

Want to learn the basics of large-scale data processing? Need to make predictive models but don’t know the right tools? This course will introduce you to open source tools you can use for parallel, distributed and scalable machine learning.

Average: 7.8 (12 votes)
Jul 17th 2017

Data analysis has replaced data acquisition as the bottleneck to evidence-based decision making --- we are drowning in it. Extracting knowledge from large, heterogeneous, and noisy datasets requires not only powerful computing resources, but the programming abstractions to use them effectively. The abstractions that emerged in the last decade blend ideas from parallel databases, distributed systems, and programming languages to create a new class of scalable data analytics platforms that form the foundation for data science at realistic scales.

Average: 6.4 (7 votes)
Jul 3rd 2017

Manipulating big data distributed over a cluster using functional concepts is rampant in industry, and is arguably one of the first widespread industrial uses of functional ideas. This is evidenced by the popularity of MapReduce and Hadoop, and most recently Apache Spark, a fast, in-memory distributed collections framework written in Scala.

No votes yet

Self Paced

Learn how to use Hadoop technologies in Microsoft Azure HDInsight to create predictive analytics and machine learning solutions. Are you ready for big data science? In this course, learn how to implement predictive analytics solutions for big data using Apache Spark in Microsoft Azure HDInsight. You will learn how to work with Scala or Python to cleanse and transform data, build machine learning models with Spark MLlib (the machine learning library in Spark), and create real-time machine learning solutions using Spark Streaming. Plus, find out how to use R Server on Spark to work with data at scale in the R language.

Average: 7.3 (4 votes)

Learn how to use Hadoop technologies like HBase, Storm, and Apache Spark in Microsoft Azure HDInsight to create real-time analytical solutions.

Average: 7.3 (3 votes)
Oct 24th 2016

Welcome to the Cloud Computing Applications course, the second part of a two-course series designed to give you a comprehensive view on the world of Cloud Computing and Big Data! In this second course we continue Cloud Computing Applications by exploring how the Cloud opens up data analytics of huge volumes of data that are static or streamed at high velocity and represent an enormous variety of information.

Average: 7.3 (4 votes)
Sep 6th 2016

SAP HANA Vora is an in-memory query engine that plugs into the Apache Spark execution framework to provide enriched interactive analytics on data stored in Hadoop. It lets you combine Big Data with corporate data in a way that is both simple and fast.

Average: 6 (1 vote)
Jul 11th 2016

Learn the underlying principles required to develop scalable machine learning pipelines and gain hands-on experience using Apache Spark. Machine learning aims to extract knowledge from data, relying on fundamental concepts in computer science, statistics, probability and optimization.

Average: 7.8 (4 votes)

Jun 15th 2016

Learn the fundamentals and architecture of Apache Spark, the leading cluster-computing framework among professionals. Spark is rapidly becoming the compute engine of choice for big data. Spark programs are more concise and often run 10-100 times faster than Hadoop MapReduce jobs. As companies realize this, Spark developers are becoming increasingly valued.

Average: 6.8 (8 votes)