Welcome to mirror list, hosted at ThFree Co, Russian Federation.

github.com/mono/Lucene.Net.Light.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'src/core/Index/SegmentInfos.cs')
-rw-r--r--src/core/Index/SegmentInfos.cs1074
1 files changed, 1074 insertions, 0 deletions
diff --git a/src/core/Index/SegmentInfos.cs b/src/core/Index/SegmentInfos.cs
new file mode 100644
index 0000000..ca5297e
--- /dev/null
+++ b/src/core/Index/SegmentInfos.cs
@@ -0,0 +1,1074 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+using System;
+using System.Collections.Generic;
+using System.IO;
+using Lucene.Net.Support;
+using ChecksumIndexInput = Lucene.Net.Store.ChecksumIndexInput;
+using ChecksumIndexOutput = Lucene.Net.Store.ChecksumIndexOutput;
+using Directory = Lucene.Net.Store.Directory;
+using IndexInput = Lucene.Net.Store.IndexInput;
+using IndexOutput = Lucene.Net.Store.IndexOutput;
+using NoSuchDirectoryException = Lucene.Net.Store.NoSuchDirectoryException;
+
+namespace Lucene.Net.Index
+{
+
+ /// <summary> A collection of segmentInfo objects with methods for operating on
+ /// those segments in relation to the file system.
+ ///
+ /// <p/><b>NOTE:</b> This API is new and still experimental
+ /// (subject to change suddenly in the next release)<p/>
+ /// </summary>
+ [Serializable]
+ public sealed class SegmentInfos : List<SegmentInfo>, ICloneable
+ {
+ private class AnonymousClassFindSegmentsFile:FindSegmentsFile
+ {
+ private void InitBlock(SegmentInfos enclosingInstance)
+ {
+ this.enclosingInstance = enclosingInstance;
+ }
+ private SegmentInfos enclosingInstance;
+ public SegmentInfos Enclosing_Instance
+ {
+ get
+ {
+ return enclosingInstance;
+ }
+
+ }
+ internal AnonymousClassFindSegmentsFile(SegmentInfos enclosingInstance, Lucene.Net.Store.Directory Param1):base(Param1)
+ {
+ InitBlock(enclosingInstance);
+ }
+
+ public /*protected internal*/ override System.Object DoBody(System.String segmentFileName)
+ {
+ Enclosing_Instance.Read(directory, segmentFileName);
+ return null;
+ }
+ }
+ /// <summary>The file format version, a negative number. </summary>
+ /* Works since counter, the old 1st entry, is always >= 0 */
+ public const int FORMAT = - 1;
+
+ /// <summary>This format adds details used for lockless commits. It differs
+ /// slightly from the previous format in that file names
+ /// are never re-used (write once). Instead, each file is
+ /// written to the next generation. For example,
+ /// segments_1, segments_2, etc. This allows us to not use
+ /// a commit lock. See <a
+ /// href="http://lucene.apache.org/java/docs/fileformats.html">file
+ /// formats</a> for details.
+ /// </summary>
+ public const int FORMAT_LOCKLESS = - 2;
+
+ /// <summary>This format adds a "hasSingleNormFile" flag into each segment info.
+ /// See <a href="http://issues.apache.org/jira/browse/LUCENE-756">LUCENE-756</a>
+ /// for details.
+ /// </summary>
+ public const int FORMAT_SINGLE_NORM_FILE = - 3;
+
+ /// <summary>This format allows multiple segments to share a single
+ /// vectors and stored fields file.
+ /// </summary>
+ public const int FORMAT_SHARED_DOC_STORE = - 4;
+
+ /// <summary>This format adds a checksum at the end of the file to
+ /// ensure all bytes were successfully written.
+ /// </summary>
+ public const int FORMAT_CHECKSUM = - 5;
+
+ /// <summary>This format adds the deletion count for each segment.
+ /// This way IndexWriter can efficiently report numDocs().
+ /// </summary>
+ public const int FORMAT_DEL_COUNT = - 6;
+
+ /// <summary>This format adds the boolean hasProx to record if any
+ /// fields in the segment store prox information (ie, have
+ /// omitTermFreqAndPositions==false)
+ /// </summary>
+ public const int FORMAT_HAS_PROX = - 7;
+
+ /// <summary>This format adds optional commit userData (String) storage. </summary>
+ public const int FORMAT_USER_DATA = - 8;
+
+ /// <summary>This format adds optional per-segment String
+ /// dianostics storage, and switches userData to Map
+ /// </summary>
+ public const int FORMAT_DIAGNOSTICS = - 9;
+
+ /* This must always point to the most recent file format. */
+ internal static readonly int CURRENT_FORMAT = FORMAT_DIAGNOSTICS;
+
+ public int counter = 0; // used to name new segments
+ /// <summary> counts how often the index has been changed by adding or deleting docs.
+ /// starting with the current time in milliseconds forces to create unique version numbers.
+ /// </summary>
+ private long version = (DateTime.UtcNow.Ticks / TimeSpan.TicksPerMillisecond);
+
+ private long generation = 0; // generation of the "segments_N" for the next commit
+ private long lastGeneration = 0; // generation of the "segments_N" file we last successfully read
+ // or wrote; this is normally the same as generation except if
+ // there was an IOException that had interrupted a commit
+
+ private IDictionary<string, string> userData = new HashMap<string, string>(); // Opaque Map<String, String> that user can specify during IndexWriter.commit
+
+ /// <summary> If non-null, information about loading segments_N files</summary>
+ /// <seealso cref="SetInfoStream">
+ /// </seealso>
+ private static System.IO.StreamWriter infoStream;
+
+ public SegmentInfo Info(int i)
+ {
+ return (SegmentInfo) this[i];
+ }
+
+ /// <summary> Get the generation (N) of the current segments_N file
+ /// from a list of files.
+ ///
+ /// </summary>
+ /// <param name="files">-- array of file names to check
+ /// </param>
+ public static long GetCurrentSegmentGeneration(System.String[] files)
+ {
+ if (files == null)
+ {
+ return - 1;
+ }
+ long max = - 1;
+ for (int i = 0; i < files.Length; i++)
+ {
+ System.String file = files[i];
+ if (file.StartsWith(IndexFileNames.SEGMENTS) && !file.Equals(IndexFileNames.SEGMENTS_GEN))
+ {
+ long gen = GenerationFromSegmentsFileName(file);
+ if (gen > max)
+ {
+ max = gen;
+ }
+ }
+ }
+ return max;
+ }
+
+ /// <summary> Get the generation (N) of the current segments_N file
+ /// in the directory.
+ ///
+ /// </summary>
+ /// <param name="directory">-- directory to search for the latest segments_N file
+ /// </param>
+ public static long GetCurrentSegmentGeneration(Directory directory)
+ {
+ try
+ {
+ return GetCurrentSegmentGeneration(directory.ListAll());
+ }
+ catch (NoSuchDirectoryException)
+ {
+ return - 1;
+ }
+ }
+
+ /// <summary> Get the filename of the current segments_N file
+ /// from a list of files.
+ ///
+ /// </summary>
+ /// <param name="files">-- array of file names to check
+ /// </param>
+
+ public static System.String GetCurrentSegmentFileName(System.String[] files)
+ {
+ return IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", GetCurrentSegmentGeneration(files));
+ }
+
+ /// <summary> Get the filename of the current segments_N file
+ /// in the directory.
+ ///
+ /// </summary>
+ /// <param name="directory">-- directory to search for the latest segments_N file
+ /// </param>
+ public static System.String GetCurrentSegmentFileName(Directory directory)
+ {
+ return IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", GetCurrentSegmentGeneration(directory));
+ }
+
+ /// <summary> Get the segments_N filename in use by this segment infos.</summary>
+ public System.String GetCurrentSegmentFileName()
+ {
+ return IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", lastGeneration);
+ }
+
+ /// <summary> Parse the generation off the segments file name and
+ /// return it.
+ /// </summary>
+ public static long GenerationFromSegmentsFileName(System.String fileName)
+ {
+ if (fileName.Equals(IndexFileNames.SEGMENTS))
+ {
+ return 0;
+ }
+ else if (fileName.StartsWith(IndexFileNames.SEGMENTS))
+ {
+ return Number.ToInt64(fileName.Substring(1 + IndexFileNames.SEGMENTS.Length));
+ }
+ else
+ {
+ throw new System.ArgumentException("fileName \"" + fileName + "\" is not a segments file");
+ }
+ }
+
+
+ /// <summary> Get the next segments_N filename that will be written.</summary>
+ [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+ public System.String GetNextSegmentFileName()
+ {
+ long nextGeneration;
+
+ if (generation == - 1)
+ {
+ nextGeneration = 1;
+ }
+ else
+ {
+ nextGeneration = generation + 1;
+ }
+ return IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", nextGeneration);
+ }
+
+ /// <summary> Read a particular segmentFileName. Note that this may
+ /// throw an IOException if a commit is in process.
+ ///
+ /// </summary>
+ /// <param name="directory">-- directory containing the segments file
+ /// </param>
+ /// <param name="segmentFileName">-- segment file to load
+ /// </param>
+ /// <throws> CorruptIndexException if the index is corrupt </throws>
+ /// <throws> IOException if there is a low-level IO error </throws>
+ public void Read(Directory directory, System.String segmentFileName)
+ {
+ bool success = false;
+
+ // Clear any previous segments:
+ Clear();
+
+ var input = new ChecksumIndexInput(directory.OpenInput(segmentFileName));
+
+ generation = GenerationFromSegmentsFileName(segmentFileName);
+
+ lastGeneration = generation;
+
+ try
+ {
+ int format = input.ReadInt();
+ if (format < 0)
+ {
+ // file contains explicit format info
+ // check that it is a format we can understand
+ if (format < CURRENT_FORMAT)
+ throw new CorruptIndexException("Unknown format version: " + format);
+ version = input.ReadLong(); // read version
+ counter = input.ReadInt(); // read counter
+ }
+ else
+ {
+ // file is in old format without explicit format info
+ counter = format;
+ }
+
+ for (int i = input.ReadInt(); i > 0; i--)
+ {
+ // read segmentInfos
+ Add(new SegmentInfo(directory, format, input));
+ }
+
+ if (format >= 0)
+ {
+ // in old format the version number may be at the end of the file
+ if (input.FilePointer >= input.Length())
+ version = (DateTime.UtcNow.Ticks / TimeSpan.TicksPerMillisecond);
+ // old file format without version number
+ else
+ version = input.ReadLong(); // read version
+ }
+
+ if (format <= FORMAT_USER_DATA)
+ {
+ if (format <= FORMAT_DIAGNOSTICS)
+ {
+ userData = input.ReadStringStringMap();
+ }
+ else if (0 != input.ReadByte())
+ {
+ // TODO: Should be read-only map
+ userData = new HashMap<string,string> {{"userData", input.ReadString()}};
+ }
+ else
+ {
+ // TODO: Should be empty read-only map
+ userData = new HashMap<string, string>();
+ }
+ }
+ else
+ {
+ // TODO: Should be empty read-only map
+ userData = new HashMap<string, string>();
+ }
+
+ if (format <= FORMAT_CHECKSUM)
+ {
+ long checksumNow = input.Checksum;
+ long checksumThen = input.ReadLong();
+ if (checksumNow != checksumThen)
+ throw new CorruptIndexException("checksum mismatch in segments file");
+ }
+ success = true;
+ }
+ finally
+ {
+ input.Close();
+ if (!success)
+ {
+ // Clear any segment infos we had loaded so we
+ // have a clean slate on retry:
+ Clear();
+ }
+ }
+ }
+
+ /// <summary> This version of read uses the retry logic (for lock-less
+ /// commits) to find the right segments file to load.
+ /// </summary>
+ /// <throws> CorruptIndexException if the index is corrupt </throws>
+ /// <throws> IOException if there is a low-level IO error </throws>
+ public void Read(Directory directory)
+ {
+
+ generation = lastGeneration = - 1;
+
+ new AnonymousClassFindSegmentsFile(this, directory).Run();
+ }
+
+ // Only non-null after prepareCommit has been called and
+ // before finishCommit is called
+ internal ChecksumIndexOutput pendingSegnOutput;
+
+ private void Write(Directory directory)
+ {
+
+ System.String segmentFileName = GetNextSegmentFileName();
+
+ // Always advance the generation on write:
+ if (generation == - 1)
+ {
+ generation = 1;
+ }
+ else
+ {
+ generation++;
+ }
+
+ var segnOutput = new ChecksumIndexOutput(directory.CreateOutput(segmentFileName));
+
+ bool success = false;
+
+ try
+ {
+ segnOutput.WriteInt(CURRENT_FORMAT); // write FORMAT
+ segnOutput.WriteLong(++version); // every write changes
+ // the index
+ segnOutput.WriteInt(counter); // write counter
+ segnOutput.WriteInt(Count); // write infos
+ for (int i = 0; i < Count; i++)
+ {
+ Info(i).Write(segnOutput);
+ }
+ segnOutput.WriteStringStringMap(userData);
+ segnOutput.PrepareCommit();
+ success = true;
+ pendingSegnOutput = segnOutput;
+ }
+ finally
+ {
+ if (!success)
+ {
+ // We hit an exception above; try to close the file
+ // but suppress any exception:
+ try
+ {
+ segnOutput.Close();
+ }
+ catch (System.Exception)
+ {
+ // Suppress so we keep throwing the original exception
+ }
+ try
+ {
+ // Try not to leave a truncated segments_N file in
+ // the index:
+ directory.DeleteFile(segmentFileName);
+ }
+ catch (System.Exception)
+ {
+ // Suppress so we keep throwing the original exception
+ }
+ }
+ }
+ }
+
+ /// <summary> Returns a copy of this instance, also copying each
+ /// SegmentInfo.
+ /// </summary>
+
+ public System.Object Clone()
+ {
+ SegmentInfos sis = new SegmentInfos();
+ for (int i = 0; i < this.Count; i++)
+ {
+ sis.Add((SegmentInfo)this[i].Clone());
+ }
+ sis.counter = this.counter;
+ sis.generation = this.generation;
+ sis.lastGeneration = this.lastGeneration;
+ // sis.pendingSegnOutput = this.pendingSegnOutput; // {{Aroush-2.9}} needed?
+ sis.userData = new HashMap<string, string>(userData);
+ sis.version = this.version;
+ return sis;
+ }
+
+ /// <summary> version number when this SegmentInfos was generated.</summary>
+ public long Version
+ {
+ get { return version; }
+ }
+
+ public long Generation
+ {
+ get { return generation; }
+ }
+
+ public long LastGeneration
+ {
+ get { return lastGeneration; }
+ }
+
+ /// <summary> Current version number from segments file.</summary>
+ /// <throws> CorruptIndexException if the index is corrupt </throws>
+ /// <throws> IOException if there is a low-level IO error </throws>
+ public static long ReadCurrentVersion(Directory directory)
+ {
+ // Fully read the segments file: this ensures that it's
+ // completely written so that if
+ // IndexWriter.prepareCommit has been called (but not
+ // yet commit), then the reader will still see itself as
+ // current:
+ var sis = new SegmentInfos();
+ sis.Read(directory);
+ return sis.version;
+ //return (long) ((System.Int64) new AnonymousClassFindSegmentsFile1(directory).Run());
+ //DIGY: AnonymousClassFindSegmentsFile1 can safely be deleted
+ }
+
+ /// <summary> Returns userData from latest segments file</summary>
+ /// <throws> CorruptIndexException if the index is corrupt </throws>
+ /// <throws> IOException if there is a low-level IO error </throws>
+ public static System.Collections.Generic.IDictionary<string, string> ReadCurrentUserData(Directory directory)
+ {
+ var sis = new SegmentInfos();
+ sis.Read(directory);
+ return sis.UserData;
+ }
+
+ /// <summary>If non-null, information about retries when loading
+ /// the segments file will be printed to this.
+ /// </summary>
+ public static void SetInfoStream(System.IO.StreamWriter infoStream)
+ {
+ SegmentInfos.infoStream = infoStream;
+ }
+
+ /* Advanced configuration of retry logic in loading
+ segments_N file */
+ private static int defaultGenFileRetryCount = 10;
+ private static int defaultGenFileRetryPauseMsec = 50;
+ private static int defaultGenLookaheadCount = 10;
+
+ /// <summary> Advanced: Gets or sets how many times to try loading the
+ /// segments.gen file contents to determine current segment
+ /// generation. This file is only referenced when the
+ /// primary method (listing the directory) fails.
+ /// </summary>
+ public static int DefaultGenFileRetryCount
+ {
+ get { return defaultGenFileRetryCount; }
+ set { defaultGenFileRetryCount = value; }
+ }
+
+ public static int DefaultGenFileRetryPauseMsec
+ {
+ set { defaultGenFileRetryPauseMsec = value; }
+ get { return defaultGenFileRetryPauseMsec; }
+ }
+
+ /// <summary> Advanced: set how many times to try incrementing the
+ /// gen when loading the segments file. This only runs if
+ /// the primary (listing directory) and secondary (opening
+ /// segments.gen file) methods fail to find the segments
+ /// file.
+ /// </summary>
+ public static int DefaultGenLookaheadCount
+ {
+ set { defaultGenLookaheadCount = value; }
+ get { return defaultGenLookaheadCount; }
+ }
+
+ /// <seealso cref="SetInfoStream">
+ /// </seealso>
+ public static StreamWriter InfoStream
+ {
+ get { return infoStream; }
+ }
+
+ private static void Message(System.String message)
+ {
+ if (infoStream != null)
+ {
+ infoStream.WriteLine("SIS [" + ThreadClass.Current().Name + "]: " + message);
+ }
+ }
+
+ /// <summary> Utility class for executing code that needs to do
+ /// something with the current segments file. This is
+ /// necessary with lock-less commits because from the time
+ /// you locate the current segments file name, until you
+ /// actually open it, read its contents, or check modified
+ /// time, etc., it could have been deleted due to a writer
+ /// commit finishing.
+ /// </summary>
+ public abstract class FindSegmentsFile
+ {
+
+ internal Directory directory;
+
+ protected FindSegmentsFile(Directory directory)
+ {
+ this.directory = directory;
+ }
+
+ public System.Object Run()
+ {
+ return Run(null);
+ }
+
+ public System.Object Run(IndexCommit commit)
+ {
+ if (commit != null)
+ {
+ if (directory != commit.Directory)
+ throw new System.IO.IOException("the specified commit does not match the specified Directory");
+ return DoBody(commit.SegmentsFileName);
+ }
+
+ System.String segmentFileName = null;
+ long lastGen = - 1;
+ long gen = 0;
+ int genLookaheadCount = 0;
+ System.IO.IOException exc = null;
+ bool retry = false;
+
+ int method = 0;
+
+ // Loop until we succeed in calling doBody() without
+ // hitting an IOException. An IOException most likely
+ // means a commit was in process and has finished, in
+ // the time it took us to load the now-old infos files
+ // (and segments files). It's also possible it's a
+ // true error (corrupt index). To distinguish these,
+ // on each retry we must see "forward progress" on
+ // which generation we are trying to load. If we
+ // don't, then the original error is real and we throw
+ // it.
+
+ // We have three methods for determining the current
+ // generation. We try the first two in parallel, and
+ // fall back to the third when necessary.
+
+ while (true)
+ {
+
+ if (0 == method)
+ {
+
+ // Method 1: list the directory and use the highest
+ // segments_N file. This method works well as long
+ // as there is no stale caching on the directory
+ // contents (NOTE: NFS clients often have such stale
+ // caching):
+ System.String[] files = null;
+
+ long genA = - 1;
+
+ files = directory.ListAll();
+
+ if (files != null)
+ genA = Lucene.Net.Index.SegmentInfos.GetCurrentSegmentGeneration(files);
+
+ Lucene.Net.Index.SegmentInfos.Message("directory listing genA=" + genA);
+
+ // Method 2: open segments.gen and read its
+ // contents. Then we take the larger of the two
+ // gens. This way, if either approach is hitting
+ // a stale cache (NFS) we have a better chance of
+ // getting the right generation.
+ long genB = - 1;
+ for (int i = 0; i < Lucene.Net.Index.SegmentInfos.defaultGenFileRetryCount; i++)
+ {
+ IndexInput genInput = null;
+ try
+ {
+ genInput = directory.OpenInput(IndexFileNames.SEGMENTS_GEN);
+ }
+ catch (System.IO.FileNotFoundException e)
+ {
+ Lucene.Net.Index.SegmentInfos.Message("segments.gen open: FileNotFoundException " + e);
+ break;
+ }
+ catch (System.IO.IOException e)
+ {
+ Lucene.Net.Index.SegmentInfos.Message("segments.gen open: IOException " + e);
+ }
+
+ if (genInput != null)
+ {
+ try
+ {
+ int version = genInput.ReadInt();
+ if (version == Lucene.Net.Index.SegmentInfos.FORMAT_LOCKLESS)
+ {
+ long gen0 = genInput.ReadLong();
+ long gen1 = genInput.ReadLong();
+ Lucene.Net.Index.SegmentInfos.Message("fallback check: " + gen0 + "; " + gen1);
+ if (gen0 == gen1)
+ {
+ // The file is consistent.
+ genB = gen0;
+ break;
+ }
+ }
+ }
+ catch (System.IO.IOException)
+ {
+ // will retry
+ }
+ finally
+ {
+ genInput.Close();
+ }
+ }
+
+ System.Threading.Thread.Sleep(new TimeSpan((System.Int64) 10000 * Lucene.Net.Index.SegmentInfos.defaultGenFileRetryPauseMsec));
+
+
+ }
+
+ Lucene.Net.Index.SegmentInfos.Message(IndexFileNames.SEGMENTS_GEN + " check: genB=" + genB);
+
+ // Pick the larger of the two gen's:
+ if (genA > genB)
+ gen = genA;
+ else
+ gen = genB;
+
+ if (gen == - 1)
+ {
+ throw new System.IO.FileNotFoundException("no segments* file found in " + directory + ": files:" + string.Join(" ", files));
+ }
+ }
+
+ // Third method (fallback if first & second methods
+ // are not reliable): since both directory cache and
+ // file contents cache seem to be stale, just
+ // advance the generation.
+ if (1 == method || (0 == method && lastGen == gen && retry))
+ {
+
+ method = 1;
+
+ if (genLookaheadCount < Lucene.Net.Index.SegmentInfos.defaultGenLookaheadCount)
+ {
+ gen++;
+ genLookaheadCount++;
+ Lucene.Net.Index.SegmentInfos.Message("look ahead increment gen to " + gen);
+ }
+ }
+
+ if (lastGen == gen)
+ {
+
+ // This means we're about to try the same
+ // segments_N last tried. This is allowed,
+ // exactly once, because writer could have been in
+ // the process of writing segments_N last time.
+
+ if (retry)
+ {
+ // OK, we've tried the same segments_N file
+ // twice in a row, so this must be a real
+ // error. We throw the original exception we
+ // got.
+ throw exc;
+ }
+
+ retry = true;
+ }
+ else if (0 == method)
+ {
+ // Segment file has advanced since our last loop, so
+ // reset retry:
+ retry = false;
+ }
+
+ lastGen = gen;
+
+ segmentFileName = IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", gen);
+
+ try
+ {
+ System.Object v = DoBody(segmentFileName);
+ Lucene.Net.Index.SegmentInfos.Message("success on " + segmentFileName);
+
+ return v;
+ }
+ catch (System.IO.IOException err)
+ {
+
+ // Save the original root cause:
+ if (exc == null)
+ {
+ exc = err;
+ }
+
+ Lucene.Net.Index.SegmentInfos.Message("primary Exception on '" + segmentFileName + "': " + err + "'; will retry: retry=" + retry + "; gen = " + gen);
+
+ if (!retry && gen > 1)
+ {
+
+ // This is our first time trying this segments
+ // file (because retry is false), and, there is
+ // possibly a segments_(N-1) (because gen > 1).
+ // So, check if the segments_(N-1) exists and
+ // try it if so:
+ System.String prevSegmentFileName = IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", gen - 1);
+
+ bool prevExists;
+ prevExists = directory.FileExists(prevSegmentFileName);
+
+ if (prevExists)
+ {
+ Lucene.Net.Index.SegmentInfos.Message("fallback to prior segment file '" + prevSegmentFileName + "'");
+ try
+ {
+ System.Object v = DoBody(prevSegmentFileName);
+ if (exc != null)
+ {
+ Lucene.Net.Index.SegmentInfos.Message("success on fallback " + prevSegmentFileName);
+ }
+ return v;
+ }
+ catch (System.IO.IOException err2)
+ {
+ Lucene.Net.Index.SegmentInfos.Message("secondary Exception on '" + prevSegmentFileName + "': " + err2 + "'; will retry");
+ }
+ }
+ }
+ }
+ }
+ }
+
+ /// <summary> Subclass must implement this. The assumption is an
+ /// IOException will be thrown if something goes wrong
+ /// during the processing that could have been caused by
+ /// a writer committing.
+ /// </summary>
+ public /*internal*/ abstract System.Object DoBody(System.String segmentFileName);
+ }
+
+ /// <summary> Returns a new SegmentInfos containg the SegmentInfo
+ /// instances in the specified range first (inclusive) to
+ /// last (exclusive), so total number of segments returned
+ /// is last-first.
+ /// </summary>
+ public SegmentInfos Range(int first, int last)
+ {
+ SegmentInfos infos = new SegmentInfos();
+ infos.AddRange(this.GetRange(first, last - first));
+ return infos;
+ }
+
+ // Carry over generation numbers from another SegmentInfos
+ internal void UpdateGeneration(SegmentInfos other)
+ {
+ lastGeneration = other.lastGeneration;
+ generation = other.generation;
+ version = other.version;
+ }
+
+ internal void RollbackCommit(Directory dir)
+ {
+ if (pendingSegnOutput != null)
+ {
+ try
+ {
+ pendingSegnOutput.Close();
+ }
+ catch (System.Exception)
+ {
+ // Suppress so we keep throwing the original exception
+ // in our caller
+ }
+
+ // Must carefully compute fileName from "generation"
+ // since lastGeneration isn't incremented:
+ try
+ {
+ System.String segmentFileName = IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", generation);
+ dir.DeleteFile(segmentFileName);
+ }
+ catch (System.Exception)
+ {
+ // Suppress so we keep throwing the original exception
+ // in our caller
+ }
+ pendingSegnOutput = null;
+ }
+ }
+
+ /// <summary>Call this to start a commit. This writes the new
+ /// segments file, but writes an invalid checksum at the
+ /// end, so that it is not visible to readers. Once this
+ /// is called you must call <see cref="FinishCommit" /> to complete
+ /// the commit or <see cref="RollbackCommit" /> to abort it.
+ /// </summary>
+ internal void PrepareCommit(Directory dir)
+ {
+ if (pendingSegnOutput != null)
+ throw new System.SystemException("prepareCommit was already called");
+ Write(dir);
+ }
+
+ /// <summary>Returns all file names referenced by SegmentInfo
+ /// instances matching the provided Directory (ie files
+ /// associated with any "external" segments are skipped).
+ /// The returned collection is recomputed on each
+ /// invocation.
+ /// </summary>
+ public System.Collections.Generic.ICollection<string> Files(Directory dir, bool includeSegmentsFile)
+ {
+ System.Collections.Generic.HashSet<string> files = new System.Collections.Generic.HashSet<string>();
+ if (includeSegmentsFile)
+ {
+ files.Add(GetCurrentSegmentFileName());
+ }
+ int size = Count;
+ for (int i = 0; i < size; i++)
+ {
+ SegmentInfo info = Info(i);
+ if (info.dir == dir)
+ {
+ files.UnionWith(Info(i).Files());
+ }
+ }
+ return files;
+ }
+
+ internal void FinishCommit(Directory dir)
+ {
+ if (pendingSegnOutput == null)
+ throw new System.SystemException("prepareCommit was not called");
+ bool success = false;
+ try
+ {
+ pendingSegnOutput.FinishCommit();
+ pendingSegnOutput.Close();
+ pendingSegnOutput = null;
+ success = true;
+ }
+ finally
+ {
+ if (!success)
+ RollbackCommit(dir);
+ }
+
+ // NOTE: if we crash here, we have left a segments_N
+ // file in the directory in a possibly corrupt state (if
+ // some bytes made it to stable storage and others
+ // didn't). But, the segments_N file includes checksum
+ // at the end, which should catch this case. So when a
+ // reader tries to read it, it will throw a
+ // CorruptIndexException, which should cause the retry
+ // logic in SegmentInfos to kick in and load the last
+ // good (previous) segments_N-1 file.
+
+ System.String fileName = IndexFileNames.FileNameFromGeneration(IndexFileNames.SEGMENTS, "", generation);
+ success = false;
+ try
+ {
+ dir.Sync(fileName);
+ success = true;
+ }
+ finally
+ {
+ if (!success)
+ {
+ try
+ {
+ dir.DeleteFile(fileName);
+ }
+ catch (System.Exception)
+ {
+ // Suppress so we keep throwing the original exception
+ }
+ }
+ }
+
+ lastGeneration = generation;
+
+ try
+ {
+ IndexOutput genOutput = dir.CreateOutput(IndexFileNames.SEGMENTS_GEN);
+ try
+ {
+ genOutput.WriteInt(FORMAT_LOCKLESS);
+ genOutput.WriteLong(generation);
+ genOutput.WriteLong(generation);
+ }
+ finally
+ {
+ genOutput.Close();
+ }
+ }
+ catch (System.Exception)
+ {
+ // It's OK if we fail to write this file since it's
+ // used only as one of the retry fallbacks.
+ }
+ }
+
+ /// <summary>Writes &amp; syncs to the Directory dir, taking care to
+ /// remove the segments file on exception
+ /// </summary>
+ public /*internal*/ void Commit(Directory dir)
+ {
+ PrepareCommit(dir);
+ FinishCommit(dir);
+ }
+
+ public System.String SegString(Directory directory)
+ {
+ lock (this)
+ {
+ var buffer = new System.Text.StringBuilder();
+ int count = Count;
+ for (int i = 0; i < count; i++)
+ {
+ if (i > 0)
+ {
+ buffer.Append(' ');
+ }
+ SegmentInfo info = Info(i);
+ buffer.Append(info.SegString(directory));
+ if (info.dir != directory)
+ buffer.Append("**");
+ }
+ return buffer.ToString();
+ }
+ }
+
+ public IDictionary<string, string> UserData
+ {
+ get { return userData; }
+ internal set {
+ userData = value ?? new HashMap<string, string>();
+ }
+ }
+
+ /// <summary>Replaces all segments in this instance, but keeps
+ /// generation, version, counter so that future commits
+ /// remain write once.
+ /// </summary>
+ internal void Replace(SegmentInfos other)
+ {
+ Clear();
+ AddRange(other);
+ lastGeneration = other.lastGeneration;
+ }
+
+ // Used only for testing
+ public bool HasExternalSegments(Directory dir)
+ {
+ int numSegments = Count;
+ for (int i = 0; i < numSegments; i++)
+ if (Info(i).dir != dir)
+ return true;
+ return false;
+ }
+
+ #region Lucene.NET (Equals & GetHashCode )
+ /// <summary>
+ /// Simple brute force implementation.
+ /// If size is equal, compare items one by one.
+ /// </summary>
+ /// <param name="obj">SegmentInfos object to check equality for</param>
+ /// <returns>true if lists are equal, false otherwise</returns>
+ public override bool Equals(object obj)
+ {
+ if (obj == null) return false;
+
+ var objToCompare = obj as SegmentInfos;
+ if (objToCompare == null) return false;
+
+ if (this.Count != objToCompare.Count) return false;
+
+ for (int idx = 0; idx < this.Count; idx++)
+ {
+ if (!this[idx].Equals(objToCompare[idx])) return false;
+ }
+
+ return true;
+ }
+
+ /// <summary>
+ /// Calculate hash code of SegmentInfos
+ /// </summary>
+ /// <returns>hash code as in java version of ArrayList</returns>
+ public override int GetHashCode()
+ {
+ int h = 1;
+ for (int i = 0; i < this.Count; i++)
+ {
+ SegmentInfo si = (this[i] as SegmentInfo);
+ h = 31 * h + (si == null ? 0 : si.GetHashCode());
+ }
+
+ return h;
+ }
+ #endregion
+ }
+} \ No newline at end of file