You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by rm...@apache.org on 2012/08/01 20:58:21 UTC
svn commit: r1368194 [20/20] - in /lucene/dev/branches/branch_4x: ./ lucene/
lucene/analysis/ lucene/analysis/common/src/java/org/tartarus/snowball/ext/
lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/
Modified: lucene/dev/branches/branch_4x/lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowball.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/branch_4x/lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowball.java?rev=1368194&r1=1368193&r2=1368194&view=diff
==============================================================================
--- lucene/dev/branches/branch_4x/lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowball.java (original)
+++ lucene/dev/branches/branch_4x/lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowball.java Wed Aug 1 18:58:20 2012
@@ -22,8 +22,10 @@ import java.io.Reader;
import org.apache.lucene.analysis.BaseTokenStreamTestCase;
import org.apache.lucene.analysis.Analyzer;
+import org.apache.lucene.analysis.MockTokenizer;
import org.apache.lucene.analysis.Tokenizer;
import org.apache.lucene.analysis.TokenStream;
+import org.apache.lucene.analysis.Analyzer.TokenStreamComponents;
import org.apache.lucene.analysis.core.KeywordTokenizer;
import org.apache.lucene.analysis.standard.StandardAnalyzer;
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
@@ -162,4 +164,21 @@ public class TestSnowball extends BaseTo
checkOneTermReuse(a, "", "");
}
}
+
+ public void testRandomStrings() throws IOException {
+ for (String lang : SNOWBALL_LANGS) {
+ checkRandomStrings(lang);
+ }
+ }
+
+ public void checkRandomStrings(final String snowballLanguage) throws IOException {
+ Analyzer a = new Analyzer() {
+ @Override
+ protected TokenStreamComponents createComponents(String fieldName, Reader reader) {
+ Tokenizer t = new MockTokenizer(reader);
+ return new TokenStreamComponents(t, new SnowballFilter(t, snowballLanguage));
+ }
+ };
+ checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER);
+ }
}
\ No newline at end of file