You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pinot.apache.org by GitBox <gi...@apache.org> on 2020/11/18 19:16:30 UTC

[GitHub] [incubator-pinot] dongxiaoman opened a new issue #6276: Spark worker timing out when processing large segment input files

dongxiaoman opened a new issue #6276:
URL: https://github.com/apache/incubator-pinot/issues/6276


   When we do `CreateSegment` on a very large segment parquet file (e.g., ~ 3GB parquet), the single worker may take more than 16 minutes collecting Stats, and the worker are actually being considered "time out" by Spark and killed.
   
   This needs to be improved if we are doing a large segment file


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pinot.apache.org
For additional commands, e-mail: commits-help@pinot.apache.org