leegee at gmail
Aug 3, 2012, 6:37 AM
Post #3 of 4
On 03/08/2012 14:56, Robert Muir wrote:
> On Wed, Aug 1, 2012 at 7:16 AM, Lee Goddard <leegee [at] gmail> wrote:
>> New to Lucene development, though I have been an indexing user for some
>> years, I find a need to develop an analyzer that reads a bespoke-format
>> (binary) file. I was wondering:
> Hello: usually you would not process such a binary file with an
> analyzer, you would parse the binary file into the Fields you care
> about and then add them to your Document.
> The analyzer is separate from that "parsing", its the way you specify
> text preprocessing at both index and query time like lowercasing,
> stemming, etc.
>> * Are there tutorials on analyzer development, or (ideally) an example
>> custom simple analyzer?
> Start with http://lucene.apache.org/core/3_6_1/api/core/org/apache/lucene/analysis/package-summary.html#package_description
>> * Is it possible to send the output of one analyzer to another, and if so,
>> is it possible to have that chain defined in the configuration of Lucene (or
>> Solr...), or would it need to be hard-coded?
> you can configure your analysis chain declaratively in Solr in a
> configuration file.
Thanks very much, Robert. And now I see the package summary JavaDoc you
pointed to, I feel quite silly.