Data Ingestion & Streaming

Splice Machine supports data from several sources. These tutorials demonstrate how to import your own data into Splice Machine, as well as how to stream data with Kafka, Apache Storm and MQTT Spark

Setting up S3 Bucket Access

Splice Machine can access S3 buckets, making it easy for you to store and manage your data on AWS. This tutorial walks you through the required steps to configure your AWS controls to allow access to S3 buckets

Importing TPCH Data

This tutorial walks you through importing importing TPCH data from our AWS bucket into your Splice Machine database and includes the SQL source for each of the TPCH queries, so you can quickly run any of them against your newly imported data

Importing Your Data

This tutorial describes how to import data into your Splice Machine database, and includes a number of examples. It also provides specific tips to help you with the details of getting your data correctly imported

Streaming Data with Kafka – Creating a Producer

This video tutorial demonstrates how to create a Kafka Producer to feed data into Splice Machine, which will subsequently be used in other tutorials

Streaming Data with Kafka – to Splice

This video tutorial demonstrates how to create a Kafka Feed, which puts messages on a Kafka Queue. This class will be used in other tutorials

Using Apache Storm with Splice Machine

This tutorial demonstrates the building and running of two different examples of integrating Storm with Splice Machine: Inserting Random Values in Splice Machine with Storm and Inserting Data into Splice Machine from MySQL

MQTT Spark Streaming

This tutorial demonstrates how to use MQTT Spark streaming with Splice Machine. MQTT is a lightweight, publish-subscribe messaging protocol designed for connecting remotely when a small footprint is required. MQTT is frequently used for data collection with the Internet of Things (IoT)