A place to get started with Apache Spark Ecosystem Components with 101 hands-on tutorial which will help you to understand the concepts of Apache Spark Ecosystem Components in detail. Note: 101 hands-on tutorial is developed using Apache Spark with Python API which is PySpark(Python programming language).
Apache Spark APIs are available in Scala, Python, Java and R programming languages. Programmers from Scala, Python, Java and R can easily develop data pipeline, machine learning model using Apache Spark APIs.
Apache Spark ecosystem components/libraries are,
Happy Learning !!!
Apache Spark APIs are available in Scala, Python, Java and R programming languages. Programmers from Scala, Python, Java and R can easily develop data pipeline, machine learning model using Apache Spark APIs.
Apache Spark
Apache Spark is a unified analytics engine for large-scale data processing.Apache Spark ecosystem components/libraries are,
- Spark Core API(RDD)
- Spark SQL(SQL, DataFrame)
- Spark Streaming, Spark Structured Streaming
- MLlib/Spark ML(Machine Learning)
- GraphX
PySpark 101 Tutorial
Happy Learning !!!
0 Comments