Skip to content

Commit a0578d6

Browse files
authored
BREAKING: De-nest AttributeSource.AttributeFactory, #1145 (#1147)
1 parent 52ee6bb commit a0578d6

File tree

63 files changed

+413
-404
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

63 files changed

+413
-404
lines changed

src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicLetterTokenizer.cs

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -41,14 +41,14 @@ namespace Lucene.Net.Analysis.Ar
4141
/// </list>
4242
/// </para>
4343
/// </summary>
44-
/// @deprecated (3.1) Use <see cref="Standard.StandardTokenizer"/> instead.
44+
/// @deprecated (3.1) Use <see cref="Standard.StandardTokenizer"/> instead.
4545
[Obsolete("(3.1) Use StandardTokenizer instead.")]
4646
public class ArabicLetterTokenizer : LetterTokenizer
4747
{
4848
/// <summary>
4949
/// Construct a new ArabicLetterTokenizer. </summary>
5050
/// <param name="matchVersion"> <see cref="LuceneVersion"/>
51-
/// to match
51+
/// to match
5252
/// </param>
5353
/// <param name="in">
5454
/// the input to split up into tokens </param>
@@ -59,7 +59,7 @@ public ArabicLetterTokenizer(LuceneVersion matchVersion, TextReader @in)
5959

6060
/// <summary>
6161
/// Construct a new <see cref="ArabicLetterTokenizer"/> using a given
62-
/// <see cref="AttributeSource.AttributeFactory"/>.
62+
/// <see cref="AttributeFactory"/>.
6363
/// </summary>
6464
/// <param name="matchVersion">
6565
/// Lucene version to match - See
@@ -82,4 +82,4 @@ protected override bool IsTokenChar(int c)
8282
return base.IsTokenChar(c) || Character.GetType(c) == UnicodeCategory.NonSpacingMark;
8383
}
8484
}
85-
}
85+
}

src/Lucene.Net.Analysis.Common/Analysis/Ar/ArabicLetterTokenizerFactory.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ namespace Lucene.Net.Analysis.Ar
2727
/// <summary>
2828
/// Factory for <see cref="ArabicLetterTokenizer"/> </summary>
2929
/// @deprecated (3.1) Use StandardTokenizerFactory instead.
30-
///
30+
///
3131
[Obsolete("(3.1) Use StandardTokenizerFactory instead.")]
3232
public class ArabicLetterTokenizerFactory : TokenizerFactory
3333
{
@@ -44,9 +44,9 @@ public ArabicLetterTokenizerFactory(IDictionary<string, string> args)
4444
}
4545
}
4646

47-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
47+
public override Tokenizer Create(AttributeFactory factory, TextReader input)
4848
{
4949
return new ArabicLetterTokenizer(m_luceneMatchVersion, factory, input);
5050
}
5151
}
52-
}
52+
}

src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKTokenizer.cs

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,7 @@
11
// Lucene version compatibility level 4.8.1
22
using J2N;
33
using Lucene.Net.Analysis.TokenAttributes;
4+
using Lucene.Net.Util;
45
using System;
56
using System.IO;
67
using System.Text.RegularExpressions;
@@ -26,7 +27,7 @@ namespace Lucene.Net.Analysis.Cjk
2627

2728
/// <summary>
2829
/// CJKTokenizer is designed for Chinese, Japanese, and Korean languages.
29-
/// <para>
30+
/// <para>
3031
/// The tokens returned are every two adjacent characters with overlap match.
3132
/// </para>
3233
/// <para>
@@ -42,7 +43,7 @@ namespace Lucene.Net.Analysis.Cjk
4243
/// please search <a
4344
/// href="http://www.google.com/search?q=word+chinese+segment">google</a>
4445
/// </summary>
45-
/// @deprecated Use StandardTokenizer, CJKWidthFilter, CJKBigramFilter, and LowerCaseFilter instead.
46+
/// @deprecated Use StandardTokenizer, CJKWidthFilter, CJKBigramFilter, and LowerCaseFilter instead.
4647
[Obsolete("Use StandardTokenizer, CJKWidthFilter, CJKBigramFilter, and LowerCaseFilter instead.")]
4748
public sealed class CJKTokenizer : Tokenizer
4849
{
@@ -96,7 +97,7 @@ public sealed class CJKTokenizer : Tokenizer
9697
private int dataLen = 0;
9798

9899
/// <summary>
99-
/// character buffer, store the characters which are used to compose
100+
/// character buffer, store the characters which are used to compose
100101
/// the returned Token
101102
/// </summary>
102103
private readonly char[] buffer = new char[MAX_WORD_LEN];
@@ -163,14 +164,14 @@ public override bool IncrementToken()
163164
{
164165
ClearAttributes();
165166

166-
// how many character(s) has been stored in buffer
167+
// how many character(s) has been stored in buffer
167168

168169
while (true) // loop until we find a non-empty token
169170
{
170171

171172
int length = 0;
172173

173-
// the position used to create Token
174+
// the position used to create Token
174175
int start = offset;
175176

176177
while (true) // loop until we've found a full token
@@ -373,4 +374,4 @@ public override void Reset()
373374
tokenType = WORD_TYPE;
374375
}
375376
}
376-
}
377+
}

src/Lucene.Net.Analysis.Common/Analysis/Cjk/CJKTokenizerFactory.cs

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ namespace Lucene.Net.Analysis.Cjk
2626

2727

2828
/// <summary>
29-
/// Factory for <see cref="CJKTokenizer"/>.
29+
/// Factory for <see cref="CJKTokenizer"/>.
3030
/// <code>
3131
/// &lt;fieldType name="text_cjk" class="solr.TextField" positionIncrementGap="100"&gt;
3232
/// &lt;analyzer&gt;
@@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.Cjk
3535
/// &lt;/fieldType&gt;
3636
/// </code>
3737
/// </summary>
38-
/// @deprecated Use <see cref="CJKBigramFilterFactory"/> instead.
38+
/// @deprecated Use <see cref="CJKBigramFilterFactory"/> instead.
3939
[Obsolete("Use CJKBigramFilterFactory instead.")]
4040
public class CJKTokenizerFactory : TokenizerFactory
4141
{
@@ -49,9 +49,9 @@ public CJKTokenizerFactory(IDictionary<string, string> args) : base(args)
4949
}
5050
}
5151

52-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader @in)
52+
public override Tokenizer Create(AttributeFactory factory, TextReader @in)
5353
{
5454
return new CJKTokenizer(factory, @in);
5555
}
5656
}
57-
}
57+
}

src/Lucene.Net.Analysis.Common/Analysis/Cn/ChineseTokenizer.cs

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,6 @@
11
// Lucene version compatibility level 4.8.1
22
using Lucene.Net.Analysis.TokenAttributes;
3+
using Lucene.Net.Util;
34
using System;
45
using System.Globalization;
56
using System.IO;
@@ -25,7 +26,7 @@ namespace Lucene.Net.Analysis.Cn
2526

2627
/// <summary>
2728
/// Tokenize Chinese text as individual chinese characters.
28-
///
29+
///
2930
/// <para>
3031
/// The difference between <see cref="ChineseTokenizer"/> and
3132
/// <see cref="Cjk.CJKTokenizer"/> is that they have different
@@ -46,10 +47,10 @@ namespace Lucene.Net.Analysis.Cn
4647
/// The problem is that when searching for C1, C1C2, C1C3,
4748
/// C4C2, C1C2C3 ... the <see cref="ChineseTokenizer"/> works, but the
4849
/// <see cref="Cjk.CJKTokenizer"/> will not work.
49-
/// </para>
50+
/// </para>
5051
/// </summary>
5152
/// @deprecated (3.1) Use <see cref="Standard.StandardTokenizer"/> instead, which has the same functionality.
52-
/// This filter will be removed in Lucene 5.0
53+
/// This filter will be removed in Lucene 5.0
5354
[Obsolete("(3.1) Use StandardTokenizer instead, which has the same functionality.")]
5455
public sealed class ChineseTokenizer : Tokenizer
5556
{
@@ -186,4 +187,4 @@ public override void Reset()
186187
offset = bufferIndex = dataLen = 0;
187188
}
188189
}
189-
}
190+
}

src/Lucene.Net.Analysis.Common/Analysis/Cn/ChineseTokenizerFactory.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ namespace Lucene.Net.Analysis.Cn
2727
/// <summary>
2828
/// Factory for <see cref="ChineseTokenizer"/>
2929
/// </summary>
30-
/// @deprecated Use <see cref="Standard.StandardTokenizerFactory"/> instead.
30+
/// @deprecated Use <see cref="Standard.StandardTokenizerFactory"/> instead.
3131
[Obsolete("Use StandardTokenizerFactory instead.")]
3232
public class ChineseTokenizerFactory : TokenizerFactory
3333
{
@@ -42,9 +42,9 @@ public ChineseTokenizerFactory(IDictionary<string, string> args)
4242
}
4343
}
4444

45-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
45+
public override Tokenizer Create(AttributeFactory factory, TextReader input)
4646
{
4747
return new ChineseTokenizer(factory, input);
4848
}
4949
}
50-
}
50+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordTokenizer.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ public KeywordTokenizer(TextReader input, int bufferSize)
5555
termAtt.ResizeBuffer(bufferSize);
5656
}
5757

58-
public KeywordTokenizer(AttributeSource.AttributeFactory factory, TextReader input, int bufferSize)
58+
public KeywordTokenizer(AttributeFactory factory, TextReader input, int bufferSize)
5959
: base(factory, input)
6060
{
6161
termAtt = AddAttribute<ICharTermAttribute>();
@@ -100,7 +100,7 @@ public override sealed bool IncrementToken()
100100
public override sealed void End()
101101
{
102102
base.End();
103-
// set final offset
103+
// set final offset
104104
offsetAtt.SetOffset(finalOffset, finalOffset);
105105
}
106106

@@ -110,4 +110,4 @@ public override void Reset()
110110
this.done = false;
111111
}
112112
}
113-
}
113+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordTokenizerFactory.cs

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -25,14 +25,14 @@ namespace Lucene.Net.Analysis.Core
2525
*/
2626

2727
/// <summary>
28-
/// Factory for <see cref="KeywordTokenizer"/>.
28+
/// Factory for <see cref="KeywordTokenizer"/>.
2929
/// <code>
3030
/// &lt;fieldType name="text_keyword" class="solr.TextField" positionIncrementGap="100"&gt;
3131
/// &lt;analyzer&gt;
3232
/// &lt;tokenizer class="solr.KeywordTokenizerFactory"/&gt;
3333
/// &lt;/analyzer&gt;
3434
/// &lt;/fieldType&gt;
35-
/// </code>
35+
/// </code>
3636
/// </summary>
3737
public class KeywordTokenizerFactory : TokenizerFactory
3838
{
@@ -47,9 +47,9 @@ public KeywordTokenizerFactory(IDictionary<string, string> args)
4747
}
4848
}
4949

50-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
50+
public override Tokenizer Create(AttributeFactory factory, TextReader input)
5151
{
5252
return new KeywordTokenizer(factory, input, KeywordTokenizer.DEFAULT_BUFFER_SIZE);
5353
}
5454
}
55-
}
55+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/LetterTokenizer.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -57,15 +57,15 @@ public LetterTokenizer(LuceneVersion matchVersion, TextReader @in)
5757

5858
/// <summary>
5959
/// Construct a new <see cref="LetterTokenizer"/> using a given
60-
/// <see cref="AttributeSource.AttributeFactory"/>.
60+
/// <see cref="AttributeFactory"/>.
6161
/// </summary>
6262
/// <param name="matchVersion">
6363
/// <see cref="LuceneVersion"/> to match</param>
6464
/// <param name="factory">
6565
/// the attribute factory to use for this <see cref="Tokenizer"/> </param>
6666
/// <param name="in">
6767
/// the input to split up into tokens </param>
68-
public LetterTokenizer(LuceneVersion matchVersion, AttributeSource.AttributeFactory factory, TextReader @in)
68+
public LetterTokenizer(LuceneVersion matchVersion, AttributeFactory factory, TextReader @in)
6969
: base(matchVersion, factory, @in)
7070
{
7171
}
@@ -79,4 +79,4 @@ protected override bool IsTokenChar(int c)
7979
return Character.IsLetter(c);
8080
}
8181
}
82-
}
82+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/LetterTokenizerFactory.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ namespace Lucene.Net.Analysis.Core
2525
*/
2626

2727
/// <summary>
28-
/// Factory for <see cref="LetterTokenizer"/>.
28+
/// Factory for <see cref="LetterTokenizer"/>.
2929
/// <code>
3030
/// &lt;fieldType name="text_letter" class="solr.TextField" positionIncrementGap="100"&gt;
3131
/// &lt;analyzer&gt;
@@ -48,9 +48,9 @@ public LetterTokenizerFactory(IDictionary<string, string> args)
4848
}
4949
}
5050

51-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
51+
public override Tokenizer Create(AttributeFactory factory, TextReader input)
5252
{
5353
return new LetterTokenizer(m_luceneMatchVersion, factory, input);
5454
}
5555
}
56-
}
56+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/LowerCaseTokenizer.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -60,15 +60,15 @@ public LowerCaseTokenizer(LuceneVersion matchVersion, TextReader @in)
6060

6161
/// <summary>
6262
/// Construct a new <see cref="LowerCaseTokenizer"/> using a given
63-
/// <see cref="AttributeSource.AttributeFactory"/>.
63+
/// <see cref="AttributeFactory"/>.
6464
/// </summary>
6565
/// <param name="matchVersion">
6666
/// <see cref="LuceneVersion"/> to match </param>
6767
/// <param name="factory">
6868
/// the attribute factory to use for this <see cref="Tokenizer"/> </param>
6969
/// <param name="in">
7070
/// the input to split up into tokens </param>
71-
public LowerCaseTokenizer(LuceneVersion matchVersion, AttributeSource.AttributeFactory factory, TextReader @in)
71+
public LowerCaseTokenizer(LuceneVersion matchVersion, AttributeFactory factory, TextReader @in)
7272
: base(matchVersion, factory, @in)
7373
{
7474
}
@@ -82,4 +82,4 @@ protected override int Normalize(int c)
8282
return Character.ToLower(c, CultureInfo.InvariantCulture); // LUCENENET specific - need to use invariant culture to match Java
8383
}
8484
}
85-
}
85+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/LowerCaseTokenizerFactory.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ namespace Lucene.Net.Analysis.Core
2525
*/
2626

2727
/// <summary>
28-
/// Factory for <see cref="LowerCaseTokenizer"/>.
28+
/// Factory for <see cref="LowerCaseTokenizer"/>.
2929
/// <code>
3030
/// &lt;fieldType name="text_lwrcase" class="solr.TextField" positionIncrementGap="100"&gt;
3131
/// &lt;analyzer&gt;
@@ -48,7 +48,7 @@ public LowerCaseTokenizerFactory(IDictionary<string, string> args)
4848
}
4949
}
5050

51-
public override Tokenizer Create(AttributeSource.AttributeFactory factory, TextReader input)
51+
public override Tokenizer Create(AttributeFactory factory, TextReader input)
5252
{
5353
return new LowerCaseTokenizer(m_luceneMatchVersion, factory, input);
5454
}
@@ -58,4 +58,4 @@ public virtual AbstractAnalysisFactory GetMultiTermComponent()
5858
return new LowerCaseFilterFactory(new Dictionary<string, string>(OriginalArgs));
5959
}
6060
}
61-
}
61+
}

src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceTokenizer.cs

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ namespace Lucene.Net.Analysis.Core
3838
/// </summary>
3939
public sealed class WhitespaceTokenizer : CharTokenizer
4040
{
41-
/// Construct a new <see cref="WhitespaceTokenizer"/>.
41+
/// Construct a new <see cref="WhitespaceTokenizer"/>.
4242
/// <param name="matchVersion"> <see cref="LuceneVersion"/> to match</param>
4343
/// <param name="in">
4444
/// the input to split up into tokens </param>
@@ -49,7 +49,7 @@ public WhitespaceTokenizer(LuceneVersion matchVersion, TextReader @in)
4949

5050
/// <summary>
5151
/// Construct a new <see cref="WhitespaceTokenizer"/> using a given
52-
/// <see cref="AttributeSource.AttributeFactory"/>.
52+
/// <see cref="AttributeFactory"/>.
5353
/// </summary>
5454
/// <param name="matchVersion"><see cref="LuceneVersion"/> to match</param>
5555
/// <param name="factory">
@@ -70,4 +70,4 @@ protected override bool IsTokenChar(int c)
7070
return !Character.IsWhiteSpace(c);
7171
}
7272
}
73-
}
73+
}

0 commit comments

Comments
 (0)