Atelier4-4 Spark
Atelier4-4 Spark
https://itabacademy.com/bigdata/hadoop/Spark/taxistreams.py
https://itabacademy.com/bigdata/hadoop/Spark/ss-test.scala
https://itabacademy.com/bigdata/hadoop/Spark/nyctaxi100.csv
# coding: utf-8
import socket
import time
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
s.bind(("localhost", 7777))
s.listen(1)
print("Started...")
while(1):
c, address = s.accept()
for row in open("nyctaxi100.csv"):
print(row)
c.send(row.encode())
time.sleep(0.5)
c.close()
import org.apache.log4j.Logger
import org.apache.log4j.Level
Logger.getLogger("org").setLevel(Level.OFF)
Logger.getLogger("akka").setLevel(Level.OFF)
import org.apache.spark._
import org.apache.spark.streaming._
import org.apache.spark.streaming.StreamingContext._
val ssc = new StreamingContext(sc, Seconds(1))
pass.print()
ssc.start()
ssc.awaitTermination()
---------------------------------------------------------------------------
[cloudera@quickstart ~]$ spark-shell
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel).
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/lib/zookeeper/lib/slf4j-log4j12-
1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/lib/flume-ng/lib/slf4j-log4j12-
1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/lib/parquet/lib/slf4j-log4j12-
1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/lib/avro/avro-tools-1.7.6-
cdh5.12.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/___/ .__/\_,_/_/ /_/\_\ version 1.6.0
/_/
Using Scala version 2.10.5 (Java HotSpot(TM) 64-Bit Server VM, Java
1.7.0_67)
Type in expressions to have them evaluated.
Type :help for more information.
22/11/07 10:22:22 WARN util.NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable
Spark context available as sc (master = local[*], app id = local-
1667845345596).
22/11/07 10:22:29 WARN shortcircuit.DomainSocketFactory: The short-circuit
local reads feature cannot be used because libhadoop cannot be loaded.
SQL context available as sqlContext.
scala> Logger.getLogger("org").setLevel(Level.OFF)
scala> Logger.getLogger("akka").setLevel(Level.OFF)
scala> import org.apache.spark._
import org.apache.spark._
scala>
scala>
scala> pass.print()
scala>
scala> ssc.start()
scala> ssc.awaitTermination()
Résultat de Traitement :