Apache Bigtop is a way for developers to try out the Apache Hadoop ecosystem quickly.
Bigtop can include components like:
Bigtop is similar to other sandboxes like Hortonworks’ data platform:
For a sandbox environment, head over to here:
You can then build sandbox images using different os (e.g. ubuntu) and w/ components like hdfs, yarn, and spark:
./build.sh -a bigtop -o ubuntu-16.04 -c "hdfs, yarn, spark"
Run your sandbox with:
docker run -d -p 50070:50070 -p 8088:8088 -p 8080:8080
bigtop/sandbox:ubuntu-16.04_hdfs_yarn_spark
Head over to the provisioner section and select your setup (e.g. Docker, Vagrant). Here you can create your clusters.
You can create and destroy clusters or run commands on individual containers.
./docker-hadoop.sh --create
3
./docker-hadoop.sh --destroy
./docker-hadoop.sh --exec 1 bash
./docker-hadoop.sh --exec
2 hadoop fs -ls /
The BigPetStore is an example application for the Hadoop/Spark ecosystems. BigPetStore generates and analyzes synthetic transaction data for a fictional chain of petstores.
The general idea is that this Spark application: