You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by GitBox <gi...@apache.org> on 2022/07/28 16:04:38 UTC

[GitHub] [hudi] glory9211 commented on issue #6194: [SUPPORT] repair deduplicate unable to find `_hoodie_record_key` in data

glory9211 commented on issue #6194:
URL: https://github.com/apache/hudi/issues/6194#issuecomment-1198346968

   > @KnightChess I added `spark.sql.parquet.mergeSchema true` to the spark properties file, then reconnected to the hudi-cli and re-tried the repair command again. The result was the same:
   > 
   > org.apache.spark.sql.AnalysisException: cannot resolve '_hoodie_record_key' given input columns: []
   > 
   > I am pretty confident that there is no schema diff in the data files.
   
   Can you verify if the duplicates are in the same file? hoodie_file column has the same value or different value
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@hudi.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org