Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Test review E-I, #259 #1014

Merged
merged 6 commits into from
Nov 12, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 3 additions & 3 deletions src/Lucene.Net.TestFramework/Support/TestFramework/Assert.cs
Original file line number Diff line number Diff line change
Expand Up @@ -930,7 +930,7 @@ public static void IsTrue(bool condition)
[MethodImpl(MethodImplOptions.AggressiveInlining)]
public static void NotNull(object anObject)
{
if (!(anObject is null))
if (anObject is not null)
_NUnit.Assert.NotNull(anObject);
}
//
Expand Down Expand Up @@ -973,7 +973,7 @@ public static void NotNull(object anObject, string message, params object[] args
[MethodImpl(MethodImplOptions.AggressiveInlining)]
public static void Null(object anObject, string message, params object[] args)
{
if (!(anObject is null))
if (anObject is not null)
_NUnit.Assert.Null(anObject, message, args);
}
//
Expand All @@ -988,7 +988,7 @@ public static void Null(object anObject, string message, params object[] args)
[MethodImpl(MethodImplOptions.AggressiveInlining)]
public static void Null(object anObject)
{
if (!(anObject is null))
if (anObject is not null)
_NUnit.Assert.Null(anObject);
}

Expand Down
4 changes: 2 additions & 2 deletions src/Lucene.Net.Tests/Index/BinaryTokenStream.cs
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ namespace Lucene.Net.Index
/// A binary tokenstream that lets you index a single
/// binary token (BytesRef value).
/// </summary>
/// <seealso> cref= CannedBinaryTokenStream </seealso>
/// <seealso cref="Lucene.Net.Analysis.CannedBinaryTokenStream" />
public sealed class BinaryTokenStream : TokenStream
{
private readonly IByteTermAttribute bytesAtt;// = addAttribute(typeof(ByteTermAttribute));
Expand Down Expand Up @@ -91,4 +91,4 @@ public override void CopyTo(IAttribute target)
}
}
}
}
}
31 changes: 16 additions & 15 deletions src/Lucene.Net.Tests/Index/Test2BBinaryDocValues.cs
Original file line number Diff line number Diff line change
Expand Up @@ -49,13 +49,14 @@ public virtual void TestFixedBinary()
{
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}
var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);
IndexWriter w = new IndexWriter(dir, config);

IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

Document doc = new Document();
var bytes = new byte[4];
Expand Down Expand Up @@ -116,13 +117,13 @@ public virtual void TestVariableBinary()
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}

var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);
IndexWriter w = new IndexWriter(dir, config);
IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

Document doc = new Document();
var bytes = new byte[4];
Expand Down Expand Up @@ -172,4 +173,4 @@ public virtual void TestVariableBinary()
dir.Dispose();
}
}
}
}
4 changes: 2 additions & 2 deletions src/Lucene.Net.Tests/Index/Test2BDocs.cs
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ public virtual void TestOverflow()
Arrays.Fill(subReaders, ir);
try
{
new MultiReader(subReaders);
_ = new MultiReader(subReaders); // LUCENENET-specific: discard result
Assert.Fail();
}
catch (Exception expected) when (expected.IsIllegalArgumentException())
Expand Down Expand Up @@ -97,4 +97,4 @@ public virtual void TestExactlyAtLimit()
dir2.Dispose();
}
}
}
}
11 changes: 6 additions & 5 deletions src/Lucene.Net.Tests/Index/Test2BPositions.cs
Original file line number Diff line number Diff line change
Expand Up @@ -42,8 +42,8 @@ namespace Lucene.Net.Index
[SuppressCodecs("SimpleText", "Memory", "Direct")]
[TestFixture]
public class Test2BPositions : LuceneTestCase
// uses lots of space and takes a few minutes
{
// uses lots of space and takes a few minutes
[Ignore("Very slow. Enable manually by removing Ignore.")]
[Test]
public virtual void Test()
Expand Down Expand Up @@ -75,7 +75,7 @@ public virtual void Test()
Field field = new Field("field", new MyTokenStream(), ft);
doc.Add(field);

int numDocs = (int.MaxValue / 26) + 1;
const int numDocs = (int.MaxValue / 26) + 1;
for (int i = 0; i < numDocs; i++)
{
w.AddDocument(doc);
Expand All @@ -91,10 +91,11 @@ public virtual void Test()

public sealed class MyTokenStream : TokenStream
{
internal readonly ICharTermAttribute termAtt;
internal readonly IPositionIncrementAttribute posIncAtt;
private readonly ICharTermAttribute termAtt;
private readonly IPositionIncrementAttribute posIncAtt;
internal int index;

// LUCENENET-specific: must call AddAttribute from ctor in .NET
public MyTokenStream()
{
termAtt = AddAttribute<ICharTermAttribute>();
Expand All @@ -121,4 +122,4 @@ public override void Reset()
}
}
}
}
}
21 changes: 11 additions & 10 deletions src/Lucene.Net.Tests/Index/Test2BPostings.cs
Original file line number Diff line number Diff line change
Expand Up @@ -54,14 +54,14 @@ public virtual void Test()
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}

var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);
var iwc = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);

IndexWriter w = new IndexWriter(dir, config);
IndexWriter w = new IndexWriter(dir, iwc);

MergePolicy mp = w.Config.MergePolicy;
if (mp is LogByteSizeMergePolicy)
Expand All @@ -77,7 +77,7 @@ public virtual void Test()
Field field = new Field("field", new MyTokenStream(), ft);
doc.Add(field);

int numDocs = (int.MaxValue / 26) + 1;
const int numDocs = (int.MaxValue / 26) + 1;
for (int i = 0; i < numDocs; i++)
{
w.AddDocument(doc);
Expand All @@ -93,9 +93,10 @@ public virtual void Test()

public sealed class MyTokenStream : TokenStream
{
internal readonly ICharTermAttribute termAtt;
private readonly ICharTermAttribute termAtt;
internal int index;

// LUCENENET-specific: must call AddAttribute from ctor in .NET
public MyTokenStream()
{
termAtt = AddAttribute<ICharTermAttribute>();
Expand All @@ -119,4 +120,4 @@ public override void Reset()
}
}
}
}
}
31 changes: 17 additions & 14 deletions src/Lucene.Net.Tests/Index/Test2BPostingsBytes.cs
Original file line number Diff line number Diff line change
Expand Up @@ -41,13 +41,13 @@ namespace Lucene.Net.Index
/// so you get > Integer.MAX_VALUE postings data for the term
/// @lucene.experimental
/// </summary>
// disable Lucene3x: older lucene formats always had this issue.
[SuppressCodecs("SimpleText", "Memory", "Direct", "Lucene3x")]
[TestFixture]
public class Test2BPostingsBytes : LuceneTestCase
// disable Lucene3x: older lucene formats always had this issue.
// @Absurd @Ignore takes ~20GB-30GB of space and 10 minutes.
// with some codecs needs more heap space as well.
{
// @Absurd @Ignore takes ~20GB-30GB of space and 10 minutes.
// with some codecs needs more heap space as well.
[Ignore("Very slow. Enable manually by removing Ignore.")]
[Test]
public virtual void Test()
Expand All @@ -58,13 +58,13 @@ public virtual void Test()
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}

var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);
IndexWriter w = new IndexWriter(dir, config);
IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

MergePolicy mp = w.Config.MergePolicy;
if (mp is LogByteSizeMergePolicy)
Expand Down Expand Up @@ -106,7 +106,8 @@ public virtual void Test()
{
((MockDirectoryWrapper)dir2).Throttling = Throttling.NEVER;
}
IndexWriter w2 = new IndexWriter(dir2, new IndexWriterConfig(TEST_VERSION_CURRENT, null));
IndexWriter w2 = new IndexWriter(dir2,
new IndexWriterConfig(TEST_VERSION_CURRENT, null));
w2.AddIndexes(mr);
w2.ForceMerge(1);
w2.Dispose();
Expand All @@ -121,7 +122,8 @@ public virtual void Test()
{
((MockDirectoryWrapper)dir3).Throttling = Throttling.NEVER;
}
IndexWriter w3 = new IndexWriter(dir3, new IndexWriterConfig(TEST_VERSION_CURRENT, null));
IndexWriter w3 = new IndexWriter(dir3,
new IndexWriterConfig(TEST_VERSION_CURRENT, null));
w3.AddIndexes(mr);
w3.ForceMerge(1);
w3.Dispose();
Expand All @@ -134,10 +136,11 @@ public virtual void Test()

public sealed class MyTokenStream : TokenStream
{
internal readonly ICharTermAttribute termAtt;
private readonly ICharTermAttribute termAtt;
internal int index;
internal int n;

// LUCENENET-specific: must call AddAttribute from ctor in .NET
public MyTokenStream()
{
termAtt = AddAttribute<ICharTermAttribute>();
Expand All @@ -162,4 +165,4 @@ public override void Reset()
}
}
}
}
}
29 changes: 15 additions & 14 deletions src/Lucene.Net.Tests/Index/Test2BSortedDocValues.cs
Original file line number Diff line number Diff line change
Expand Up @@ -48,12 +48,13 @@ public virtual void TestFixedSorted()
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}

IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));
IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

Document doc = new Document();
var bytes = new byte[2];
Expand Down Expand Up @@ -110,13 +111,13 @@ public virtual void Test2BOrds()
((MockDirectoryWrapper)dir).Throttling = Throttling.NEVER;
}

var config = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE);
IndexWriter w = new IndexWriter(dir, config);
IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

Document doc = new Document();
var bytes = new byte[4];
Expand Down Expand Up @@ -169,4 +170,4 @@ public virtual void Test2BOrds()

// TODO: variable
}
}
}
31 changes: 16 additions & 15 deletions src/Lucene.Net.Tests/Index/Test2BTerms.cs
Original file line number Diff line number Diff line change
Expand Up @@ -59,12 +59,12 @@ public class Test2BTerms : LuceneTestCase

private sealed class MyTokenStream : TokenStream
{
internal readonly int tokensPerDoc;
internal int tokenCount;
private readonly int tokensPerDoc;
private int tokenCount;
public readonly IList<BytesRef> savedTerms = new JCG.List<BytesRef>();
internal int nextSave;
internal long termCounter;
internal readonly Random random;
private int nextSave;
private long termCounter;
private readonly Random random;

public MyTokenStream(Random random, int tokensPerDoc)
: base(new MyAttributeFactory(AttributeFactory.DEFAULT_ATTRIBUTE_FACTORY))
Expand Down Expand Up @@ -140,7 +140,7 @@ public override object Clone()

private sealed class MyAttributeFactory : AttributeFactory
{
internal readonly AttributeFactory @delegate;
private readonly AttributeFactory @delegate;

public MyAttributeFactory(AttributeFactory @delegate)
{
Expand Down Expand Up @@ -172,7 +172,7 @@ public virtual void Test2BTerms_Mem()
throw RuntimeException.Create("this test cannot run with PreFlex codec");
}
Console.WriteLine("Starting Test2B");
long TERM_COUNT = ((long)int.MaxValue) + 100000000;
const long TERM_COUNT = ((long)int.MaxValue) + 100000000;

int TERMS_PER_DOC = TestUtil.NextInt32(Random, 100000, 1000000);

Expand All @@ -188,12 +188,13 @@ public virtual void Test2BTerms_Mem()

if (true)
{
IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));
IndexWriter w = new IndexWriter(dir,
new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(Random))
.SetMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH)
.SetRAMBufferSizeMB(256.0)
.SetMergeScheduler(new ConcurrentMergeScheduler())
.SetMergePolicy(NewLogMergePolicy(false, 10))
.SetOpenMode(OpenMode.CREATE));

MergePolicy mp = w.Config.MergePolicy;
if (mp is LogByteSizeMergePolicy)
Expand All @@ -202,7 +203,7 @@ public virtual void Test2BTerms_Mem()
((LogByteSizeMergePolicy)mp).MaxMergeMB = 1024 * 1024 * 1024;
}

Documents.Document doc = new Documents.Document();
Document doc = new Document();
MyTokenStream ts = new MyTokenStream(Random, TERMS_PER_DOC);

FieldType customType = new FieldType(TextField.TYPE_NOT_STORED);
Expand Down Expand Up @@ -311,4 +312,4 @@ private void TestSavedTerms(IndexReader r, IList<BytesRef> terms)
Assert.IsFalse(failed);
}
}
}
}
Loading
Loading