Hortonworks Sandbox for HDP and HDF is your chance to get started on learning, developing, testing and trying out new features. Each download comes preconfigured with interactive tutorials, sample data and developments from the Apache community.

The table stacks have been raised. Now, your minimal viable enterprise system is a combination of Big Data and related technologies. The minimum infrastructure is now hybrid and spans multiple clouds and devices on-premise. You need, at a minimum, a standard open-source Hadoop platform like HDP 2.5 or ODPi with HDFS, YARN, Hive2LLAP, HBase, or Phoenix as your base for massive petabyte storage.

On top of that, you need to be running Spark 2 jobs on YARN for various Machine Learning, streaming, graph, and batch jobs. Through Spark or through containers, you need to run TensorFlow and other Deep Learning packages at scale in your massive distributed cluster.

You also need to be able to coordinate, ingest, transmit, translate and store from thousands of different devices from IIoT, SCADA, mobile, raspberry Pis, and various devices. You will need to stream from various sources in an Apache open way not tied to the cloud vendor of the day.

Thus, Kafka and MQTT are required. Getting trapped in non-locally installable technology that doesn't extend to your 1GB devices is a weakness in your environment. Basically:

Want a cheap place to run some Machine Learning, NLP, or Deep Learning on a small Raspberry Pi-sized cloud box? Read this article for the rundown of Digital Ocean, Linode, and a bunch of cheap offerings out there.

Hortonworks Sandbox for HDP and HDF is your chance to get started on learning, developing, testing and trying out new features. Each download comes preconfigured with interactive tutorials, sample data and developments from the Apache community.