GitHub / aimanamri / raspberry-pi4-hadoop-spark-cluster
This is a self-documentation of learning distributed data storage, parallel processing, and Linux OS using Apache Hadoop, Apache Spark and Raspbian OS. In this project, 3-node cluster will be setup using Raspberry Pi 4, install HDFS and run Spark processing jobs via YARN.
JSON API: http://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/aimanamri%2Fraspberry-pi4-hadoop-spark-cluster
PURL: pkg:github/aimanamri/raspberry-pi4-hadoop-spark-cluster
Stars: 1
Forks: 0
Open issues: 0
License: gpl-3.0
Language: Shell
Size: 5.21 MB
Dependencies parsed at: Pending
Created at: 11 months ago
Updated at: 7 months ago
Pushed at: 11 months ago
Last synced at: 2 months ago
Commit Stats
Commits: 13
Authors: 1
Mean commits per author: 13.0
Development Distribution Score: 0.0
More commit stats: https://commits.ecosyste.ms/hosts/GitHub/repositories/aimanamri/raspberry-pi4-hadoop-spark-cluster
Topics: big-data, distributed-storage, hadoop-cluster, hdfs, parallel-processing, pyspark, raspberry-pi-4, spark-cluster, spark-shell, yarn