Apache Spark

Apache Spark is an open-source parallel processing framework for storing and processing Big Data across clustered computers. Spark can be used to perform computations much faster than Hadoop can rather Hadoop and Spark can be used together efficiently. Spark is written in Scala, which is considered the primary language for interacting with the Spark Core engine, but it doesn’t require developers to know Scala, which executes inside a Java Virtual Machine (JVM)…




Love podcasts or audiobooks? Learn on the go with our new app.

Recommended from Medium

Setup a 3D and Video Studio on Ubuntu Linux for Free

Linux: Do you need to know Linux as a DevOps Engineer? — Shaik Wahab

Orchestrate & Build ETL pipeline using Azure Databricks and Azure Data Factory v2 (Part — 2)

Css ←→ Swift

If you want to make an app for that, stop.

[LeetCode]#1684. Count the Number of Consistent Strings

How to Batch Resize Images in Word Document

Top 5 Web Technologies to Learn

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Jagandeep Singh

Jagandeep Singh

More from Medium

A Convenient Way to Run PySpark

Introduction to Datasets in Spark

Spark Tuning, Optimization, and Performance Techniques

Managing RPC and Stack Overflow Errors from High Iterations in Spark ALS