You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@lucenenet.apache.org by George Aroush <ge...@aroush.net> on 2007/04/04 01:05:34 UTC

RE: tokenizer optimizations

Do you have some examples?  Are those stop-words you are referring to?

SQL's analyzer/tokenizer doesn't necessarily match that of Lucene.Net's.

-- George

> -----Original Message-----
> From: Michael Paine [mailto:michael@justanswer.com] 
> Sent: Wednesday, April 04, 2007 5:29 AM
> To: lucene-net-user@incubator.apache.org
> Subject: tokenizer optimizations
> 
> I have found many words not tokenized (and indexed) by 
> Lucene.NET which are
> tokenized/indexed by SQL Server 2005 full-text.   Is there a 
> way to optimize
> lucene's tokenizer so it handles more words? 
>  
>