Streaming with Apache Spark 2.0

Knoldus Inc.
Knoldus - Technical Insights
2 min readSep 30, 2016

Hello geeks we were discussed about Apache Spark 2.0 with hive in earlier blog.

Now i am going to describe how can we use spark to stream the data .

At first we need to understand this new Spark Streaming architecture .

Spark 2.0 simplified the API for Streaming and lets us to access stream data in form of DataFrame and DataSet. Hence with new architecture we can process our streamed data according to our business logic with DataFrame. This is the simple concept behind above architecture.

So here we have two approach to use Spark Streaming programmetically :

  • by using predefined receiver , and
  • by creating Custom-Receiver

At first we will stream our data using predefined receiver .

Add the following dependencies :

“org.apache.spark” %% “spark-core” % “2.0.0”,
“org.apache.spark” %% “spark-sql” % “2.0.0”,
“org.apache.spark” %% “spark-hive” % “2.0.0”,
“org.apache.spark” %% “spark-streaming” % “2.0.0”

Now as we know entry point of Spark in current version is SparkSession . So ,

[code language=”Scala”]
val sparkSession = SparkSession.builder
.master(“local”)
.appName(“demo”)
.getOrCreate()
[/code]

Now you need stream receiver :

[code language=”Scala”]
val dataFrame : DataFrame = sparkSession.read.json(“src/main/resources/dummy.json”)
[/code]

Now we get the data of stream here we can perform our any bussines logic with dataframe.

[code language=”Scala”]
dataFrame.distinct().show()
[/code]

Find Demo code here.
Continue Reading to Streaming with Custom Receiver.

Thanks

KNOLDUS-advt-sticker

--

--

Knoldus Inc.
Knoldus - Technical Insights

Group of smart Engineers with a Product mindset who partner with your business to drive competitive advantage | www.knoldus.com