You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by cloud-fan <gi...@git.apache.org> on 2017/03/11 03:33:53 UTC

[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

GitHub user cloud-fan opened a pull request:

    https://github.com/apache/spark/pull/17253

    [SPARK-19916][SQL] simplify bad file handling

    ## What changes were proposed in this pull request?
    
    We should only have one centre place to try catch the exception for corrupted files.
    
    ## How was this patch tested?
    
    existing test

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/cloud-fan/spark bad-file

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/17253.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #17253
    
----
commit 05febbdc58f566426796dbf814000381b309062f
Author: Wenchen Fan <we...@databricks.com>
Date:   2017-03-11T01:09:49Z

    simplify bad file handling

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    **[Test build #74420 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/74420/testReport)** for PR 17253 at commit [`ad64848`](https://github.com/apache/spark/commit/ad64848b2212de30bc7935e8b8cbdade3e38b7a9).
     * This patch passes all tests.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by viirya <gi...@git.apache.org>.
Github user viirya commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105525158
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -130,54 +144,35 @@ class FileScanRDD(
               // Sets InputFileBlockHolder for the file block's information
               InputFileBlockHolder.set(currentFile.filePath, currentFile.start, currentFile.length)
     
    -          try {
    -            if (ignoreCorruptFiles) {
    -              currentIterator = new NextIterator[Object] {
    -                private val internalIter = {
    -                  try {
    -                    // The readFunction may read files before consuming the iterator.
    -                    // E.g., vectorized Parquet reader.
    -                    readFunction(currentFile)
    -                  } catch {
    -                    case e @(_: RuntimeException | _: IOException) =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      Iterator.empty
    -                  }
    -                }
    -
    -                override def getNext(): AnyRef = {
    -                  try {
    -                    if (internalIter.hasNext) {
    -                      internalIter.next()
    -                    } else {
    -                      finished = true
    -                      null
    -                    }
    -                  } catch {
    -                    case e: IOException =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      finished = true
    -                      null
    +          if (ignoreCorruptFiles) {
    +            currentIterator = new NextIterator[Object] {
    +              // The readFunction may read some bytes before consuming the iterator, e.g.,
    +              // vectorized Parquet reader. Here we use lazy val to delay the creation of
    +              // iterator so that we will throw exception in `getNext`.
    +              private lazy val internalIter = readCurrentFile()
    +
    +              override def getNext(): AnyRef = {
    +                try {
    +                  if (internalIter.hasNext) {
    +                    internalIter.next()
    +                  } else {
    +                    finished = true
    +                    null
                       }
    +                } catch {
    +                  // Throw FileNotFoundException even `ignoreCorruptFiles` is true
    +                  case e: java.io.FileNotFoundException => throw e
    --- End diff --
    
    nit: `FileNotFoundException` will be thrown anyway, do wee need this case?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105587493
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormat.scala ---
    @@ -90,16 +90,14 @@ trait FileFormat {
        * @param options A set of string -> string configuration options.
        * @return
        */
    -  def buildReader(
    +  protected def buildReader(
           sparkSession: SparkSession,
           dataSchema: StructType,
           partitionSchema: StructType,
           requiredSchema: StructType,
           filters: Seq[Filter],
           options: Map[String, String],
           hadoopConf: Configuration): PartitionedFile => Iterator[InternalRow] = {
    -    // TODO: Remove this default implementation when the other formats have been ported
    --- End diff --
    
    Actually we don't need to implement this method in all sub-classes. Some `FileFormat` may override `buildReaderWithPartitionValues` directly(parquet), Some `FileFormat` may not be used in read path(HiveFileFormat)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105526080
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -130,54 +144,35 @@ class FileScanRDD(
               // Sets InputFileBlockHolder for the file block's information
               InputFileBlockHolder.set(currentFile.filePath, currentFile.start, currentFile.length)
     
    -          try {
    -            if (ignoreCorruptFiles) {
    -              currentIterator = new NextIterator[Object] {
    -                private val internalIter = {
    -                  try {
    -                    // The readFunction may read files before consuming the iterator.
    -                    // E.g., vectorized Parquet reader.
    -                    readFunction(currentFile)
    -                  } catch {
    -                    case e @(_: RuntimeException | _: IOException) =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      Iterator.empty
    -                  }
    -                }
    -
    -                override def getNext(): AnyRef = {
    -                  try {
    -                    if (internalIter.hasNext) {
    -                      internalIter.next()
    -                    } else {
    -                      finished = true
    -                      null
    -                    }
    -                  } catch {
    -                    case e: IOException =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      finished = true
    -                      null
    +          if (ignoreCorruptFiles) {
    +            currentIterator = new NextIterator[Object] {
    +              // The readFunction may read some bytes before consuming the iterator, e.g.,
    +              // vectorized Parquet reader. Here we use lazy val to delay the creation of
    +              // iterator so that we will throw exception in `getNext`.
    +              private lazy val internalIter = readCurrentFile()
    +
    +              override def getNext(): AnyRef = {
    +                try {
    +                  if (internalIter.hasNext) {
    +                    internalIter.next()
    +                  } else {
    +                    finished = true
    +                    null
                       }
    +                } catch {
    +                  // Throw FileNotFoundException even `ignoreCorruptFiles` is true
    +                  case e: java.io.FileNotFoundException => throw e
    --- End diff --
    
    `FileNotFoundException` extends `IOException`


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by viirya <gi...@git.apache.org>.
Github user viirya commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105529164
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileFormat.scala ---
    @@ -90,16 +90,14 @@ trait FileFormat {
        * @param options A set of string -> string configuration options.
        * @return
        */
    -  def buildReader(
    +  protected def buildReader(
           sparkSession: SparkSession,
           dataSchema: StructType,
           partitionSchema: StructType,
           requiredSchema: StructType,
           filters: Seq[Filter],
           options: Map[String, String],
           hadoopConf: Configuration): PartitionedFile => Iterator[InternalRow] = {
    -    // TODO: Remove this default implementation when the other formats have been ported
    --- End diff --
    
    No more TODO here?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    **[Test build #74367 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/74367/testReport)** for PR 17253 at commit [`05febbd`](https://github.com/apache/spark/commit/05febbdc58f566426796dbf814000381b309062f).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    cc @sameeragarwal @viirya


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by viirya <gi...@git.apache.org>.
Github user viirya commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105525303
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -44,7 +44,7 @@ case class PartitionedFile(
         filePath: String,
         start: Long,
         length: Long,
    -    locations: Array[String] = Array.empty) {
    +    @transient locations: Array[String] = Array.empty) {
    --- End diff --
    
    do we need to mark it as `transient`? `filePartitions: Seq[FilePartition])` is already `transient` in `FileScanRDD`.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    Merged build finished. Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by cloud-fan <gi...@git.apache.org>.
Github user cloud-fan commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105526137
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -44,7 +44,7 @@ case class PartitionedFile(
         filePath: String,
         start: Long,
         length: Long,
    -    locations: Array[String] = Array.empty) {
    +    @transient locations: Array[String] = Array.empty) {
    --- End diff --
    
    this is not for `FileScanRDD.filePartitions`, this is for `FilePartition`s that sent by scheduler. The location is only useful during planning, we should not send it to executors.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/74367/
    Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    Merged build finished. Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by viirya <gi...@git.apache.org>.
Github user viirya commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    LGTM with very minor comment.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by viirya <gi...@git.apache.org>.
Github user viirya commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    LGTM


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by apurvis <gi...@git.apache.org>.
Github user apurvis commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105552466
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -130,54 +144,35 @@ class FileScanRDD(
               // Sets InputFileBlockHolder for the file block's information
               InputFileBlockHolder.set(currentFile.filePath, currentFile.start, currentFile.length)
     
    -          try {
    -            if (ignoreCorruptFiles) {
    -              currentIterator = new NextIterator[Object] {
    -                private val internalIter = {
    -                  try {
    -                    // The readFunction may read files before consuming the iterator.
    -                    // E.g., vectorized Parquet reader.
    -                    readFunction(currentFile)
    -                  } catch {
    -                    case e @(_: RuntimeException | _: IOException) =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      Iterator.empty
    -                  }
    -                }
    -
    -                override def getNext(): AnyRef = {
    -                  try {
    -                    if (internalIter.hasNext) {
    -                      internalIter.next()
    -                    } else {
    -                      finished = true
    -                      null
    -                    }
    -                  } catch {
    -                    case e: IOException =>
    -                      logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    -                      finished = true
    -                      null
    +          if (ignoreCorruptFiles) {
    +            currentIterator = new NextIterator[Object] {
    +              // The readFunction may read some bytes before consuming the iterator, e.g.,
    +              // vectorized Parquet reader. Here we use lazy val to delay the creation of
    +              // iterator so that we will throw exception in `getNext`.
    +              private lazy val internalIter = readCurrentFile()
    +
    +              override def getNext(): AnyRef = {
    +                try {
    +                  if (internalIter.hasNext) {
    +                    internalIter.next()
    +                  } else {
    +                    finished = true
    +                    null
                       }
    +                } catch {
    +                  // Throw FileNotFoundException even `ignoreCorruptFiles` is true
    +                  case e: java.io.FileNotFoundException => throw e
    +                  case e @ (_: RuntimeException | _: IOException) =>
    +                    logWarning(s"Skipped the rest content in the corrupted file: $currentFile", e)
    --- End diff --
    
    Better English: "Skipped the rest _of the_ content in the corrupted file:"


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by asfgit <gi...@git.apache.org>.
Github user asfgit closed the pull request at:

    https://github.com/apache/spark/pull/17253


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by gatorsmile <gi...@git.apache.org>.
Github user gatorsmile commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    Thanks! Merging to master.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark pull request #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by srowen <gi...@git.apache.org>.
Github user srowen commented on a diff in the pull request:

    https://github.com/apache/spark/pull/17253#discussion_r105552491
  
    --- Diff: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/FileScanRDD.scala ---
    @@ -121,6 +121,20 @@ class FileScanRDD(
             nextElement
           }
     
    +      private def readCurrentFile(): Iterator[InternalRow] = {
    +        try {
    +          readFunction(currentFile)
    +        } catch {
    +          case e: java.io.FileNotFoundException =>
    --- End diff --
    
    Why not import the class? same below


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by AmplabJenkins <gi...@git.apache.org>.
Github user AmplabJenkins commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    Test PASSed.
    Refer to this link for build results (access rights to CI server needed): 
    https://amplab.cs.berkeley.edu/jenkins//job/SparkPullRequestBuilder/74420/
    Test PASSed.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    **[Test build #74420 has started](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/74420/testReport)** for PR 17253 at commit [`ad64848`](https://github.com/apache/spark/commit/ad64848b2212de30bc7935e8b8cbdade3e38b7a9).


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org


[GitHub] spark issue #17253: [SPARK-19916][SQL] simplify bad file handling

Posted by SparkQA <gi...@git.apache.org>.
Github user SparkQA commented on the issue:

    https://github.com/apache/spark/pull/17253
  
    **[Test build #74367 has finished](https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/74367/testReport)** for PR 17253 at commit [`05febbd`](https://github.com/apache/spark/commit/05febbdc58f566426796dbf814000381b309062f).
     * This patch passes all tests.
     * This patch merges cleanly.
     * This patch adds no public classes.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org