Limitations of Apache Spark
Dear friends, Today we will discuss about the limitations of Apache Spark and its Apache Spark disadvantages. As all we know that Apache Spark is a fast, in-memory data processing engine with elegant and expressive development APIs to allow data workers to efficiently execute Streaming, Machine Learning or SQL workloads that require fast iterative access to datasets. Below are the few features of it : 1. Apache Spark achieves high performance for both batch and streaming data, using a state-of-the-art DAG scheduler, a query optimizer, and a physical execution engine. 2. We can write applications quickly in Java, Scala, Python, R, and SQL. Spark offers over 80 high-level operators that make it easy to build parallel apps and we can use it interactively from the Scala, Python, R, and SQL shells. e.g. df = spark.read.json("json_file_name.json") df.where("age > 21") .select("name.first") .show() // Spark's Pyth