You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Ashish Kumar Singh <as...@gmail.com> on 2016/05/20 21:04:34 UTC
Logstash to collect Spark logs
We are trying to collect Spark logs using logstash for parsing app logs
and collecting useful info.
We can read the Nodemanager logs but unable to read Spark application logs
using Logstash .
Current Setup for Spark logs and Logstash
1- Spark runs on Yarn .
2- Using log4j socketAppenders to write logs to tcp port .
3- Below lines added in log4j.properties of Yarn and Spark conf:
main.logger=RFA,SA
log4j.appender.SA=org.apache.log4j.net.SocketAppender
log4j.appender.SA.Port=4560
log4j.appender.SA.RemoteHost=${hostname}
log4j.appender.SA.ReconnectionDelay=10000
log4j.appender.SA.Application=NM-${user.dir}
4-Logstash input
input {
log4j {
mode => "server"
host => "0.0.0.0"
port => 4560
type => "log4j"
}
}
Any help on reading Spark logs via Logstash will be appreciated .
Also, is there a better way to collect Spark logs via Logstash ?