Apache Spark

Apache Spark is an open-source parallel processing framework for storing and processing Big Data across clustered computers. Spark can be used to perform computations much faster than Hadoop can rather Hadoop and Spark can be used together efficiently. Spark is written in Scala, which is considered the primary language for interacting with the Spark Core engine, but it doesn’t require developers to know Scala, which executes inside a Java Virtual Machine (JVM)…

Love podcasts or audiobooks? Learn on the go with our new app.

Recommended from Medium

Lessons from my first year as a Sr. engineer

Why Am I Starting A Blog? (And the Process Behind It)

Design choices behind the MuesliSwap smart contract

Everything about the Magento 2 Store, View, and Website

Configuring HTTPD Server on Docker Container

How to Manage Azure Key Vault with Terraform

Rails Threading Gains

Creating Azure Custom Linux VM Image

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Jagandeep Singh

Jagandeep Singh

More from Medium

Installing Spark on Windows 10

Spark Dataframes

Setting up Pyspark in mac

Apache Sqoop