You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@lucenenet.apache.org by George Aroush <ge...@aroush.net> on 2007/04/04 01:05:34 UTC
RE: tokenizer optimizations
Do you have some examples? Are those stop-words you are referring to?
SQL's analyzer/tokenizer doesn't necessarily match that of Lucene.Net's.
-- George
> -----Original Message-----
> From: Michael Paine [mailto:michael@justanswer.com]
> Sent: Wednesday, April 04, 2007 5:29 AM
> To: lucene-net-user@incubator.apache.org
> Subject: tokenizer optimizations
>
> I have found many words not tokenized (and indexed) by
> Lucene.NET which are
> tokenized/indexed by SQL Server 2005 full-text. Is there a
> way to optimize
> lucene's tokenizer so it handles more words?
>
>