You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by sz...@apache.org on 2011/04/21 18:05:31 UTC
svn commit: r1095761 - in /hadoop/common/trunk: ./
src/java/org/apache/hadoop/fs/ src/java/org/apache/hadoop/fs/shell/
src/test/core/org/apache/hadoop/cli/
Author: szetszwo
Date: Thu Apr 21 16:05:30 2011
New Revision: 1095761
URL: http://svn.apache.org/viewvc?rev=1095761&view=rev
Log:
HADOOP-7233. Refactor ls to conform to new FsCommand class. Contributed by Daryn Sharp
Added:
hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Ls.java
Modified:
hadoop/common/trunk/CHANGES.txt
hadoop/common/trunk/src/java/org/apache/hadoop/fs/FsShell.java
hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Command.java
hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Count.java
hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/FsCommand.java
hadoop/common/trunk/src/test/core/org/apache/hadoop/cli/testConf.xml
Modified: hadoop/common/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/CHANGES.txt?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/CHANGES.txt (original)
+++ hadoop/common/trunk/CHANGES.txt Thu Apr 21 16:05:30 2011
@@ -96,6 +96,9 @@ Trunk (unreleased changes)
HADOOP-7230. Move "fs -help" shell command tests from HDFS to COMMOM; see
also HDFS-1844. (Daryn Sharp via szetszwo)
+ HADOOP-7233. Refactor ls to conform to new FsCommand class. (Daryn Sharp
+ via szetszwo)
+
OPTIMIZATIONS
BUG FIXES
Modified: hadoop/common/trunk/src/java/org/apache/hadoop/fs/FsShell.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/java/org/apache/hadoop/fs/FsShell.java?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/src/java/org/apache/hadoop/fs/FsShell.java (original)
+++ hadoop/common/trunk/src/java/org/apache/hadoop/fs/FsShell.java Thu Apr 21 16:05:30 2011
@@ -613,87 +613,6 @@ public class FsShell extends Configured
}
}
-
- /**
- * Get a listing of all files in that match the file pattern <i>srcf</i>.
- * @param srcf a file pattern specifying source files
- * @param recursive if need to list files in subdirs
- * @throws IOException
- * @see org.apache.hadoop.fs.FileSystem#globStatus(Path)
- */
- private int ls(String srcf, boolean recursive) throws IOException {
- Path srcPath = new Path(srcf);
- FileSystem srcFs = srcPath.getFileSystem(this.getConf());
- FileStatus[] srcs = srcFs.globStatus(srcPath);
- if (srcs==null || srcs.length==0) {
- throw new FileNotFoundException("Cannot access " + srcf +
- ": No such file or directory.");
- }
-
- boolean printHeader = (srcs.length == 1) ? true: false;
- int numOfErrors = 0;
- for(int i=0; i<srcs.length; i++) {
- numOfErrors += ls(srcs[i], srcFs, recursive, printHeader);
- }
- return numOfErrors == 0 ? 0 : -1;
- }
-
- /* list all files under the directory <i>src</i>
- * ideally we should provide "-l" option, that lists like "ls -l".
- */
- private int ls(FileStatus src, FileSystem srcFs, boolean recursive,
- boolean printHeader) throws IOException {
- final String cmd = recursive? "lsr": "ls";
- final FileStatus[] items = shellListStatus(cmd, srcFs, src);
- if (items == null) {
- return 1;
- } else {
- int numOfErrors = 0;
- if (!recursive && printHeader) {
- if (items.length != 0) {
- System.out.println("Found " + items.length + " items");
- }
- }
-
- int maxReplication = 3, maxLen = 10, maxOwner = 0,maxGroup = 0;
-
- for(int i = 0; i < items.length; i++) {
- FileStatus stat = items[i];
- int replication = String.valueOf(stat.getReplication()).length();
- int len = String.valueOf(stat.getLen()).length();
- int owner = String.valueOf(stat.getOwner()).length();
- int group = String.valueOf(stat.getGroup()).length();
-
- if (replication > maxReplication) maxReplication = replication;
- if (len > maxLen) maxLen = len;
- if (owner > maxOwner) maxOwner = owner;
- if (group > maxGroup) maxGroup = group;
- }
-
- for (int i = 0; i < items.length; i++) {
- FileStatus stat = items[i];
- Path cur = stat.getPath();
- String mdate = dateForm.format(new Date(stat.getModificationTime()));
-
- System.out.print((stat.isDirectory() ? "d" : "-") +
- stat.getPermission() + " ");
- System.out.printf("%"+ maxReplication +
- "s ", (stat.isFile() ? stat.getReplication() : "-"));
- if (maxOwner > 0)
- System.out.printf("%-"+ maxOwner + "s ", stat.getOwner());
- if (maxGroup > 0)
- System.out.printf("%-"+ maxGroup + "s ", stat.getGroup());
- System.out.printf("%"+ maxLen + "d ", stat.getLen());
- System.out.print(mdate + " ");
- System.out.println(cur.toUri().getPath());
- if (recursive && stat.isDirectory()) {
- numOfErrors += ls(stat,srcFs, recursive, printHeader);
- }
- }
- return numOfErrors;
- }
- }
-
/**
* Show the size of a partition in the filesystem that contains
* the specified <i>path</i>.
@@ -1400,7 +1319,7 @@ public class FsShell extends Configured
String summary = "hadoop fs is the command to execute fs commands. " +
"The full syntax is: \n\n" +
"hadoop fs [-fs <local | file system URI>] [-conf <configuration file>]\n\t" +
- "[-D <property=value>] [-ls <path>] [-lsr <path>] [-df [<path>]] [-du [-s] [-h] <path>]\n\t" +
+ "[-D <property=value>] [-df [<path>]] [-du [-s] [-h] <path>]\n\t" +
"[-dus <path>] [-mv <src> <dst>] [-cp <src> <dst>] [-rm [-skipTrash] <src>]\n\t" +
"[-rmr [-skipTrash] <src>] [-put <localsrc> ... <dst>] [-copyFromLocal <localsrc> ... <dst>]\n\t" +
"[-moveFromLocal <localsrc> ... <dst>] [" +
@@ -1429,21 +1348,6 @@ public class FsShell extends Configured
"\t\tappear first on the command line. Exactly one additional\n" +
"\t\targument must be specified. \n";
-
- String ls = "-ls <path>: \tList the contents that match the specified file pattern. If\n" +
- "\t\tpath is not specified, the contents of /user/<currentUser>\n" +
- "\t\twill be listed. Directory entries are of the form \n" +
- "\t\t\tdirName (full path) <dir> \n" +
- "\t\tand file entries are of the form \n" +
- "\t\t\tfileName(full path) <r n> size \n" +
- "\t\twhere n is the number of replicas specified for the file \n" +
- "\t\tand size is the size of the file, in bytes.\n";
-
- String lsr = "-lsr <path>: \tRecursively list the contents that match the specified\n" +
- "\t\tfile pattern. Behaves very similarly to hadoop fs -ls,\n" +
- "\t\texcept that the data is shown for all the entries in the\n" +
- "\t\tsubtree.\n";
-
String df = "-df [<path>]: \tShows the capacity, free and used space of the filesystem.\n"+
"\t\tIf the filesystem has multiple partitions, and no path to a particular partition\n"+
"\t\tis specified, then the status of the root partitions will be shown.\n";
@@ -1573,17 +1477,13 @@ public class FsShell extends Configured
Command instance = commandFactory.getInstance("-" + cmd);
if (instance != null) {
- System.out.println(instance.getDescription());
+ printHelp(instance);
} else if ("fs".equals(cmd)) {
System.out.println(fs);
} else if ("conf".equals(cmd)) {
System.out.println(conf);
} else if ("D".equals(cmd)) {
System.out.println(D);
- } else if ("ls".equals(cmd)) {
- System.out.println(ls);
- } else if ("lsr".equals(cmd)) {
- System.out.println(lsr);
} else if ("df".equals(cmd)) {
System.out.println(df);
} else if ("du".equals(cmd)) {
@@ -1644,13 +1544,11 @@ public class FsShell extends Configured
System.out.println(summary);
for (String thisCmdName : commandFactory.getNames()) {
instance = commandFactory.getInstance(thisCmdName);
- System.out.println(instance.getUsage());
+ System.out.println("\t[" + instance.getUsage() + "]");
}
System.out.println("\t[-help [cmd]]\n");
System.out.println(fs);
- System.out.println(ls);
- System.out.println(lsr);
System.out.println(df);
System.out.println(du);
System.out.println(dus);
@@ -1678,14 +1576,29 @@ public class FsShell extends Configured
System.out.println(chgrp);
for (String thisCmdName : commandFactory.getNames()) {
- instance = commandFactory.getInstance(thisCmdName);
- System.out.println(instance.getDescription());
+ printHelp(commandFactory.getInstance(thisCmdName));
}
System.out.println(help);
}
}
+ // TODO: will eventually auto-wrap the text, but this matches the expected
+ // output for the hdfs tests...
+ private void printHelp(Command instance) {
+ boolean firstLine = true;
+ for (String line : instance.getDescription().split("\n")) {
+ String prefix;
+ if (firstLine) {
+ prefix = instance.getUsage() + ":\t";
+ firstLine = false;
+ } else {
+ prefix = "\t\t";
+ }
+ System.out.println(prefix + line);
+ }
+ }
+
/**
* Apply operation specified by 'cmd' on all parameters
* starting from argv[startindex].
@@ -1720,10 +1633,6 @@ public class FsShell extends Configured
delete(argv[i], true, rmSkipTrash);
} else if ("-df".equals(cmd)) {
df(argv[i]);
- } else if ("-ls".equals(cmd)) {
- exitCode = ls(argv[i], false);
- } else if ("-lsr".equals(cmd)) {
- exitCode = ls(argv[i], true);
} else if ("-touchz".equals(cmd)) {
touchz(argv[i]);
} else if ("-text".equals(cmd)) {
@@ -1781,8 +1690,7 @@ public class FsShell extends Configured
} else if ("-D".equals(cmd)) {
System.err.println("Usage: java FsShell" +
" [-D <[property=value>]");
- } else if ("-ls".equals(cmd) || "-lsr".equals(cmd) ||
- "-du".equals(cmd) || "-dus".equals(cmd) ||
+ } else if ("-du".equals(cmd) || "-dus".equals(cmd) ||
"-touchz".equals(cmd) || "-mkdir".equals(cmd) ||
"-text".equals(cmd)) {
System.err.println("Usage: java FsShell" +
@@ -1822,8 +1730,6 @@ public class FsShell extends Configured
System.err.println("Usage: java FsShell [" + TAIL_USAGE + "]");
} else {
System.err.println("Usage: java FsShell");
- System.err.println(" [-ls <path>]");
- System.err.println(" [-lsr <path>]");
System.err.println(" [-df [<path>]]");
System.err.println(" [-du [-s] [-h] <path>]");
System.err.println(" [-dus <path>]");
@@ -1965,18 +1871,6 @@ public class FsShell extends Configured
"-chown".equals(cmd) ||
"-chgrp".equals(cmd)) {
exitCode = FsShellPermissions.changePermissions(cmd, argv, i, this);
- } else if ("-ls".equals(cmd)) {
- if (i < argv.length) {
- exitCode = doall(cmd, argv, i);
- } else {
- exitCode = ls(Path.CUR_DIR, false);
- }
- } else if ("-lsr".equals(cmd)) {
- if (i < argv.length) {
- exitCode = doall(cmd, argv, i);
- } else {
- exitCode = ls(Path.CUR_DIR, true);
- }
} else if ("-mv".equals(cmd)) {
exitCode = rename(argv, getConf());
} else if ("-cp".equals(cmd)) {
Modified: hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Command.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Command.java?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Command.java (original)
+++ hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Command.java Thu Apr 21 16:05:30 2011
@@ -79,6 +79,14 @@ abstract public class Command extends Co
name = cmdName;
}
+ protected void setRecursive(boolean flag) {
+ recursive = flag;
+ }
+
+ protected boolean isRecursive() {
+ return recursive;
+ }
+
/**
* Execute the command on the input path
*
@@ -138,7 +146,8 @@ abstract public class Command extends Co
displayError(e);
}
- return (numErrors == 0) ? exitCode : 1;
+ // TODO: -1 should be reserved for syntax error, 1 should be failure
+ return (numErrors == 0) ? exitCode : -1;
}
/**
@@ -244,7 +253,18 @@ abstract public class Command extends Co
protected void processNonexistentPathArgument(PathData item)
throws IOException {
// TODO: this should be more posix-like: ex. "No such file or directory"
- throw new FileNotFoundException("Can not find listing for " + item);
+ throw new FileNotFoundException(getFnfText(item.path));
+ }
+
+ /**
+ * TODO: A crutch until the text is standardized across commands...
+ * Eventually an exception that takes the path as an argument will
+ * replace custom text
+ * @param path the thing that doesn't exist
+ * @returns String in printf format
+ */
+ protected String getFnfText(Path path) {
+ throw new RuntimeException(path + ": No such file or directory");
}
/**
@@ -309,7 +329,7 @@ abstract public class Command extends Co
if (stats != null && stats.length == 0) { // glob failed to match
// TODO: this should be more posix-like: ex. "No such file or directory"
- throw new FileNotFoundException("Can not find listing for " + pattern);
+ throw new FileNotFoundException(getFnfText(path));
}
List<PathData> items = new LinkedList<PathData>();
@@ -371,7 +391,9 @@ abstract public class Command extends Co
* @return "name options"
*/
public String getUsage() {
- return getCommandField("USAGE");
+ String cmd = "-" + getCommandName();
+ String usage = getCommandField("USAGE");
+ return usage.isEmpty() ? cmd : cmd + " " + usage;
}
/**
Modified: hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Count.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Count.java?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Count.java (original)
+++ hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Count.java Thu Apr 21 16:05:30 2011
@@ -26,6 +26,7 @@ import org.apache.hadoop.classification.
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.ContentSummary;
import org.apache.hadoop.fs.FsShell;
+import org.apache.hadoop.fs.Path;
/**
* Count the number of directories, files, bytes, quota, and remaining quota.
@@ -43,13 +44,13 @@ public class Count extends FsCommand {
}
public static final String NAME = "count";
- public static final String USAGE = "-" + NAME + " [-q] <path> ...";
- public static final String DESCRIPTION = CommandUtils.formatDescription(USAGE,
- "Count the number of directories, files and bytes under the paths",
- "that match the specified file pattern. The output columns are:",
- "DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME or",
- "QUOTA REMAINING_QUATA SPACE_QUOTA REMAINING_SPACE_QUOTA ",
- " DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME");
+ public static final String USAGE = "[-q] <path> ...";
+ public static final String DESCRIPTION =
+ "Count the number of directories, files and bytes under the paths\n" +
+ "that match the specified file pattern. The output columns are:\n" +
+ "DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME or\n" +
+ "QUOTA REMAINING_QUATA SPACE_QUOTA REMAINING_SPACE_QUOTA \n" +
+ " DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME";
private boolean showQuotas;
@@ -86,4 +87,10 @@ public class Count extends FsCommand {
ContentSummary summary = src.fs.getContentSummary(src.path);
out.println(summary.toString(showQuotas) + src.path);
}
+
+ // TODO: remove when the error is commonized...
+ @Override
+ protected String getFnfText(Path path) {
+ return "Can not find listing for " + path;
+ }
}
Modified: hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/FsCommand.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/FsCommand.java?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/FsCommand.java (original)
+++ hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/FsCommand.java Thu Apr 21 16:05:30 2011
@@ -42,7 +42,8 @@ abstract public class FsCommand extends
* @param factory where to register the class
*/
public static void registerCommands(CommandFactory factory) {
- Count.registerCommands(factory);
+ factory.registerCommands(Count.class);
+ factory.registerCommands(Ls.class);
}
protected FsCommand() {}
Added: hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Ls.java
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Ls.java?rev=1095761&view=auto
==============================================================================
--- hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Ls.java (added)
+++ hadoop/common/trunk/src/java/org/apache/hadoop/fs/shell/Ls.java Thu Apr 21 16:05:30 2011
@@ -0,0 +1,155 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.fs.shell;
+
+import java.io.IOException;
+import java.text.SimpleDateFormat;
+import java.util.Date;
+import java.util.LinkedList;
+
+import org.apache.hadoop.classification.InterfaceAudience;
+import org.apache.hadoop.classification.InterfaceStability;
+import org.apache.hadoop.fs.FileStatus;
+import org.apache.hadoop.fs.Path;
+
+/**
+ * Get a listing of all files in that match the file patterns.
+ */
+@InterfaceAudience.Private
+@InterfaceStability.Unstable
+
+class Ls extends FsCommand {
+ public static void registerCommands(CommandFactory factory) {
+ factory.addClass(Ls.class, "-ls");
+ factory.addClass(Lsr.class, "-lsr");
+ }
+
+ public static final String NAME = "ls";
+ public static final String USAGE = "[<path> ...]";
+ public static final String DESCRIPTION =
+ "List the contents that match the specified file pattern. If\n" +
+ "path is not specified, the contents of /user/<currentUser>\n" +
+ "will be listed. Directory entries are of the form \n" +
+ "\tdirName (full path) <dir> \n" +
+ "and file entries are of the form \n" +
+ "\tfileName(full path) <r n> size \n" +
+ "where n is the number of replicas specified for the file \n" +
+ "and size is the size of the file, in bytes.";
+
+ protected static final SimpleDateFormat dateFormat =
+ new SimpleDateFormat("yyyy-MM-dd HH:mm");
+
+ protected int maxRepl = 3, maxLen = 10, maxOwner = 0, maxGroup = 0;
+ protected String lineFormat;
+
+ @Override
+ protected void processOptions(LinkedList<String> args)
+ throws IOException {
+ CommandFormat cf = new CommandFormat(null, 0, Integer.MAX_VALUE, "R");
+ cf.parse(args);
+ setRecursive(cf.getOpt("R"));
+ if (args.isEmpty()) args.add(Path.CUR_DIR);
+ }
+
+ @Override
+ protected void processPaths(PathData parent, PathData ... items)
+ throws IOException {
+ // implicitly recurse once for cmdline directories
+ if (parent == null && items[0].stat.isDirectory()) {
+ recursePath(items[0]);
+ return;
+ }
+
+ if (!isRecursive() && items.length != 0) {
+ out.println("Found " + items.length + " items");
+ }
+ adjustColumnWidths(items);
+ super.processPaths(parent, items);
+ }
+
+ @Override
+ protected void processPath(PathData item) throws IOException {
+ FileStatus stat = item.stat;
+ String line = String.format(lineFormat,
+ (stat.isDirectory() ? "d" : "-"),
+ stat.getPermission(),
+ (stat.isFile() ? stat.getReplication() : "-"),
+ stat.getOwner(),
+ stat.getGroup(),
+ stat.getLen(),
+ dateFormat.format(new Date(stat.getModificationTime())),
+ item.path.toUri().getPath()
+ );
+ out.println(line);
+ }
+
+ /**
+ * Compute column widths and rebuild the format string
+ * @param items to find the max field width for each column
+ */
+ private void adjustColumnWidths(PathData items[]) {
+ for (PathData item : items) {
+ FileStatus stat = item.stat;
+ maxRepl = maxLength(maxRepl, stat.getReplication());
+ maxLen = maxLength(maxLen, stat.getLen());
+ maxOwner = maxLength(maxOwner, stat.getOwner());
+ maxGroup = maxLength(maxGroup, stat.getGroup());
+ }
+
+ StringBuilder fmt = new StringBuilder();
+ fmt.append("%s%s "); // permission string
+ fmt.append("%" + maxRepl + "s ");
+ fmt.append("%-" + maxOwner + "s ");
+ fmt.append("%-" + maxGroup + "s ");
+ fmt.append("%" + maxLen + "d ");
+ fmt.append("%s %s"); // mod time & path
+ lineFormat = fmt.toString();
+ }
+
+ private int maxLength(int n, Object value) {
+ return Math.max(n, (value != null) ? String.valueOf(value).length() : 0);
+ }
+
+ // TODO: remove when the error is commonized...
+ @Override
+ protected String getFnfText(Path path) {
+ return "Cannot access " + path.toUri() + ": No such file or directory.";
+ }
+
+ /**
+ * Get a recursive listing of all files in that match the file patterns.
+ * Same as "-ls -R"
+ */
+ public static class Lsr extends Ls {
+ public static final String NAME = "lsr";
+ public static final String USAGE = Ls.USAGE;
+ public static final String DESCRIPTION =
+ "Recursively list the contents that match the specified\n" +
+ "file pattern. Behaves very similarly to hadoop fs -ls,\n" +
+ "except that the data is shown for all the entries in the\n" +
+ "subtree.";
+
+ @Override
+ protected void processOptions(LinkedList<String> args)
+ throws IOException {
+ args.addFirst("-R");
+ super.processOptions(args);
+ }
+ }
+}
Modified: hadoop/common/trunk/src/test/core/org/apache/hadoop/cli/testConf.xml
URL: http://svn.apache.org/viewvc/hadoop/common/trunk/src/test/core/org/apache/hadoop/cli/testConf.xml?rev=1095761&r1=1095760&r2=1095761&view=diff
==============================================================================
--- hadoop/common/trunk/src/test/core/org/apache/hadoop/cli/testConf.xml (original)
+++ hadoop/common/trunk/src/test/core/org/apache/hadoop/cli/testConf.xml Thu Apr 21 16:05:30 2011
@@ -54,7 +54,7 @@
<comparators>
<comparator>
<type>RegexpComparator</type>
- <expected-output>^-ls <path>:( |\t)*List the contents that match the specified file pattern. If( )*</expected-output>
+ <expected-output>^-ls \[<path> \.\.\.\]:( |\t)*List the contents that match the specified file pattern. If( )*</expected-output>
</comparator>
<comparator>
<type>RegexpComparator</type>
@@ -101,7 +101,7 @@
<comparators>
<comparator>
<type>RegexpComparator</type>
- <expected-output>^-lsr <path>:( |\t)*Recursively list the contents that match the specified( )*</expected-output>
+ <expected-output>^-lsr \[<path> \.\.\.\]:( |\t)*Recursively list the contents that match the specified( )*</expected-output>
</comparator>
<comparator>
<type>RegexpComparator</type>
@@ -222,7 +222,7 @@
<comparators>
<comparator>
<type>RegexpComparator</type>
- <expected-output>^-count \[-q\] <path> \.\.\.: Count the number of directories, files and bytes under the paths( )*</expected-output>
+ <expected-output>^-count \[-q\] <path> \.\.\.:( |\t)*Count the number of directories, files and bytes under the paths( )*</expected-output>
</comparator>
<comparator>
<type>RegexpComparator</type>