EMR Spark Streaming Job Stdout logging disappears -


when launch spark streaming job on emr (cluster mode), can see stdout job first few moments disappears...

i can see few log lines @ following location in s3 (i setup emr copy logs s3 bucket): s3-us-west-1.amazonaws.com//spark/logs/j-iemn2tmesrek/containers/application_1454718762107_0001/container_1454718762107_0001_01_000001/stdout.gz

after ~10 seconds of streaming job running, no more stdout being delivered logs.

is emr redirecting stdout somewhere else?

turned out executors not getting log4j configuration.

i used bootstrap step place log4j.properties in /tmp/log4jproperties.

then using spark-submit following arguments

--conf spark.executor.extrajavaoptions=-dlog4j.configuration=/tmp/log4j.properties --files file:///tmp/log4j.properties


Comments

Popular posts from this blog

sublimetext3 - what keyboard shortcut is to comment/uncomment for this script tag in sublime -

java - No use of nillable="0" in SOAP Webservice -

ubuntu - Laravel 5.2 quickstart guide gives Not Found Error -