Lucene.Net  3.0.3
Lucene.Net is a port of the Lucene search engine library, written in C# and targeted at .NET runtime users.
 All Classes Namespaces Files Functions Variables Typedefs Enumerations Properties Pages
KeywordAnalyzer.cs
Go to the documentation of this file.
1 /*
2  * Licensed to the Apache Software Foundation (ASF) under one or more
3  * contributor license agreements. See the NOTICE file distributed with
4  * this work for additional information regarding copyright ownership.
5  * The ASF licenses this file to You under the Apache License, Version 2.0
6  * (the "License"); you may not use this file except in compliance with
7  * the License. You may obtain a copy of the License at
8  *
9  * http://www.apache.org/licenses/LICENSE-2.0
10  *
11  * Unless required by applicable law or agreed to in writing, software
12  * distributed under the License is distributed on an "AS IS" BASIS,
13  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14  * See the License for the specific language governing permissions and
15  * limitations under the License.
16  */
17 
18 namespace Lucene.Net.Analysis
19 {
20 
21  /// <summary> "Tokenizes" the entire stream as a single token. This is useful
22  /// for data like zip codes, ids, and some product names.
23  /// </summary>
25  {
26  public KeywordAnalyzer()
27  {
28  SetOverridesTokenStreamMethod<KeywordAnalyzer>();
29  }
30  public override TokenStream TokenStream(System.String fieldName, System.IO.TextReader reader)
31  {
32  return new KeywordTokenizer(reader);
33  }
34  public override TokenStream ReusableTokenStream(System.String fieldName, System.IO.TextReader reader)
35  {
36  if (overridesTokenStreamMethod)
37  {
38  // LUCENE-1678: force fallback to tokenStream() if we
39  // have been subclassed and that subclass overrides
40  // tokenStream but not reusableTokenStream
41  return TokenStream(fieldName, reader);
42  }
43  var tokenizer = (Tokenizer) PreviousTokenStream;
44  if (tokenizer == null)
45  {
46  tokenizer = new KeywordTokenizer(reader);
47  PreviousTokenStream = tokenizer;
48  }
49  else
50  tokenizer.Reset(reader);
51  return tokenizer;
52  }
53  }
54 }