Skip to content

Latest commit

 

History

History
312 lines (215 loc) · 8.57 KB

README.md

File metadata and controls

312 lines (215 loc) · 8.57 KB

Spark Streamer implementation

Using macOS.

Install dependencies:

$ brew install maven
$ brew install apache-spark

Important if you use jenv:

$ jenv enable-plugin maven
$ echo 'JAVA_HOME=$(/usr/libexec/java_home -v $(jenv version-name))' >> ~/.mavenrc

or use maven with jenv exec, for example: jenv exec mvn clean install

$ mvn package
$ spark-submit --class "zoltan.nz.App" --master local[4] target/spark-streamer-1.0-SNAPSHOT.jar

Important:

  • Use Scala v2.12

Connect to Apache Kafka on PORT 9092

Launch Kafka Producer

https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html https://databricks.com/blog/2016/07/28/structured-streaming-in-apache-spark.html https://spark.apache.org/docs/2.2.0/structured-streaming-kafka-integration.html

Add SQL Kafka maven package

Experiment #01 - Install Spark from Source

I tried the standard installation process, but Kafka and Spark did not work together as expected. Various errors blocked running of the basic examples. Issues with libraries, problem with unsolvable AbstractClasses, etc.

Instead of playing with standard installation methods, I started to build Spark from source code.

Cloning the original project

$ git clone git@github.com:apache/spark.git

Try to build

$ cd spark
$  build/mvn -DskipTests clean package

Failing with this issue:

[INFO] Using zinc server for incremental compilation
[info] 'compiler-interface' not yet compiled for Scala 2.11.8. Compiling...
error: scala.reflect.internal.MissingRequirementError: object java.lang.Object in compiler mirror not found.

A suggested solution:

Removing scala-compile-first and scala-test-compile-first from pom.xml

<execution>
  <id>scala-compile-first</id>
  <goals>
    <goal>compile</goal>
  </goals>
</execution>
<execution>
  <id>scala-test-compile-first</id>
  <goals>
    <goal>testCompile</goal>
  </goals>
</execution>

The above approach worked, I had a new build, but it was not working properly.

Found an other approach: davidB/scala-maven-plugin#185

So I added back the above compilation steps and I used the following command to run the build:

$ ./build/zinc-0.3.15/bin/zinc -shutdown
$ build/mvn -DskipTests -DrecompileMode=all clean package

Checking the build:

$ ./bin/spark-shell
> sc.parallelize(1 to 1000).count() 

We have access to Web UI: http://localhost:4040

So Experiment #01 is ready, we have a working Spark implementation.

Experiment #02 - Playing with basic examples

$ ./bin/run-example JavaSparkPi

Running basic streaming example:

Run Netcat server in a separate terminal:

$ nc -lk 9999

Run Spark's example:

$ ./bin/run-example org.apache.spark.examples.streaming.JavaCustomReceiver localhost 9999

Running the Kafka example:

Run Kafka separately.

The following is the suggested example:

$ bin/run-example streaming.JavaDirectKafkaWordCount localhost:9092 test

Running the above I get the following error:

Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/spark/streaming/kafka010/LocationStrategies

Experiment #03 - Trying to solve the above problem

No solution yet.

Experiment #04 - Trying to add different library

Following the suggested linking description: https://spark.apache.org/docs/latest/streaming-programming-guide.html#linking

I added spark-streaming-kafka-0-10 2.11 to pom.xml. Source: https://search.maven.org/#search%7Cga%7C1%7Cg%3A%22org.apache.spark%22%20AND%20v%3A%222.3.0%22

Updated the App.java. Source: https://spark.apache.org/docs/latest/streaming-kafka-integration.html Source: https://spark.apache.org/docs/latest/streaming-kafka-0-10-integration.html

On this website (https://spark.apache.org/docs/2.2.0/structured-streaming-kafka-integration.html), we can get the following example:

// Subscribe to 1 topic
DataFrame<Row> df = spark
  .readStream()
  .format("kafka")
  .option("kafka.bootstrap.servers", "host1:port1,host2:port2")
  .option("subscribe", "topic1")
  .load()
df.selectExpr("CAST(key AS STRING)", "CAST(value AS STRING)")

This is not right anymore, because in v2.3 DataFrame type is not exists anymore.

Reading the DataFrame and Datasets Guide (https://spark.apache.org/docs/latest/sql-programming-guide.html), we can get the following little hint:

In Scala and Java, a DataFrame is represented by a Dataset of Rows. In the Scala API, DataFrame is simply a type alias of Dataset[Row]. While, in Java API, users need to use Dataset to represent a DataFrame.

So we have to use Dataset<Row> to make it work.

The following code is almost working:

package nz.zoltan;

import org.apache.spark.api.java.function.FlatMapFunction;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Encoders;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession;
import org.apache.spark.sql.streaming.StreamingQuery;
import org.apache.spark.sql.streaming.StreamingQueryException;

import java.util.Arrays;

public class App {

  public static void main(String[] args) throws StreamingQueryException {

    SparkSession spark = SparkSession
      .builder()
      .appName("JavaStructuredNetworkWordCount")
      .getOrCreate();

    // Create DataFrame representing the stream of input lines from connection to localhost:9999
    Dataset<Row> df = spark
      .readStream()
      .format("kafka")
      .option("kafka.bootstrap.servers", "localhost:9092")
      .option("subscribe", "boerse.dev")
      .load();
    df.selectExpr("CAST(key AS STRING)", "CAST(value AS STRING)");

    // Split the lines into words
    Dataset<String> words = df
      .as(Encoders.STRING())
      .flatMap((FlatMapFunction<String, String>) x -> Arrays.asList(x.split(" ")).iterator(), Encoders.STRING());

    // Generate running word count
    Dataset<Row> wordCounts = words.groupBy("value").count();

    // Start running the query that prints the running counts to the console
    StreamingQuery query = wordCounts.writeStream()
      .outputMode("complete")
      .format("console")
      .start();

    query.awaitTermination();
  }
}

Compile and run:

$ mvn clean package
$ spark-submit --packages org.apache.spark:spark-sql-kafka-0-10_2.11:2.3.0 target/spark-streamer-1.0-SNAPSHOT.jar

However, the above string would fail, because we get simple text stream from Kafka.

Commenting out most of the code and changing outputMode:

  StreamingQuery query = df.writeStream()
      .outputMode("append")
      .format("console")
      .start();

Finally, the streaming starts working. :D

Experiment #05 - Debugging Spark App

spark-submit --conf spark.driver.extraJavaOptions=-agentlib:jdwp=transport=dt_socket,server=y,address=8000,suspend=n --packages org.apache.spark:spark-sql-kafka-0-10_2.11:2.3.0 target/spark-streamer-1.0-SNAPSHOT.jar

Experiment #06 - Using local

Launching Spark in a standalone mode: https://spark.apache.org/docs/latest/submitting-applications.html#master-urls

Issue: In case of error like this:

"java.lang.NoClassDefFoundError: scala/Product$class"

Have to check the Scala maven package version in pom.xml.

Issue: Cannot find Kafka package.

And don't miss to add the following dependency:

    <dependency>
      <groupId>org.apache.spark</groupId>
      <artifactId>spark-sql-kafka-0-10_2.11</artifactId>
      <version>2.3.0</version>
    </dependency>

Important!

  • For development and debugging, we have to use local mode to run our Spark session.
      .config("spark.master", "local")

Dockerization of a Maven App

Inspirations:

$ docker build -t spark-streamer .
$ docker run spark-streamer:latest mvn exec:java

Using host machine maven repository:

$docker run -v "$HOME/.m2":/root/.m2 spark-streamer:latest mvn exec:java