You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Steve Loughran (JIRA)" <ji...@apache.org> on 2014/09/25 11:31:34 UTC

[jira] [Commented] (HADOOP-11128) abstracting out the scale tests for FileSystem Contract tests

    [ https://issues.apache.org/jira/browse/HADOOP-11128?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14147575#comment-14147575 ] 

Steve Loughran commented on HADOOP-11128:
-----------------------------------------

# the sizes and even test timeout can be configured by properties in the contract-test files
# some of the over 5GB tests, which test specific situations in the object stores, may need to skip explicitly if the test machine isn't set to test the big file size.
# One test I don't think anyone has done yet (I think I may have done some of it in openstack), is the wide & deep tree; that may exhibit different problems from the wide directory and deep path.

> abstracting out the scale tests for FileSystem Contract tests
> -------------------------------------------------------------
>
>                 Key: HADOOP-11128
>                 URL: https://issues.apache.org/jira/browse/HADOOP-11128
>             Project: Hadoop Common
>          Issue Type: Improvement
>            Reporter: Juan Yu
>
> Currently we have some scale tests for openstack and s3a. For now we'll just trust HDFS to handle files >5GB and delete thousands of file in a directory properly.
> We should abstract out the scale tests so it can be applied to all FileSystems.
> A few things to consider for scale tests:
> scale tests rely on the tester having good/stable upload bandwidth, might need large disk space. It needs to be configurable or optional.
> scale tests might need long time to finish, consider have test timeout configurable if possible



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)