Lucene++ - a full-featured, c++ search engine
API Documentation
►NLucene | |
CAbstractAllTermDocs | Base class for enumerating all but deleted docs |
CAbstractField | |
CAcceptAllSinkFilter | |
CAllTermDocs | |
CAnalyzer | An Analyzer builds TokenStreams, which analyze text. It thus represents a policy for extracting index terms from text |
CArray | Utility template class to handle sharable arrays of simple data types |
CArrayData | |
CASCIIFoldingFilter | This class converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if one exists |
CAttribute | Base class for Attributes that can be added to a AttributeSource |
CAttributeFactory | |
CAttributeSource | An AttributeSource contains a list of different Attribute s, and methods to add and get them. There can only be a single instance of an attribute in the same AttributeSource instance. This is ensured by passing in the actual type of the Attribute (Class<Attribute>) to the addAttribute(Class) , which then checks if an instance of that type is already present. If yes, it returns the instance, otherwise it creates a new instance and returns it |
CAttributeSourceState | This class holds the state of an AttributeSource |
CAveragePayloadFunction | Calculate the final score as the average score of all payloads seen |
CBase64 | |
CBaseCharFilter | Base utility class for implementing a CharFilter . You subclass this, and then record mappings by calling addOffCorrectMap , and then invoke the correct method to correct an offset |
CBitSet | |
CBitUtil | A variety of high efficiency bit twiddling routines |
CBitVector | Optimized implementation of a vector of bits |
CBooleanClause | A clause in a BooleanQuery |
CBooleanQuery | A Query that matches documents matching boolean combinations of other queries, eg. TermQuery s, PhraseQuery s or other BooleanQuerys |
CBooleanScorer | BooleanScorer uses a ~16k array to score windows of docs. So it scores docs 0-16k first, then docs 16-32k, etc. For each window it iterates through all query terms and accumulates a score in table[doc%16k]. It also stores in the table a bitmask representing which terms contributed to the score. Non-zero scores are chained in a linked list. At the end of scoring each window it then iterates through the linked list and, if the bitmask matches the boolean constraints, collects a hit. For boolean queries with lots of frequent terms this can be much faster, since it does not need to update a priority queue for each posting, instead performing constant-time operations per posting. The only downside is that it results in hits being delivered out-of-order within the window, which means it cannot be nested within other scorers. But it works well as a top-level scorer |
CBooleanScorer2 | See the description in BooleanScorer, comparing BooleanScorer & BooleanScorer2 |
CBooleanScorerCollector | |
CBucket | |
CBucketScorer | |
CBucketTable | A simple hash table of document scores within a range |
CBufferedDeletes | Holds buffered deletes, by docID, term or query. We hold two instances of this class: one for the deletes prior to the last flush, the other for deletes after the last flush. This is so if we need to abort (discard all buffered docs) we can also discard the buffered deletes yet keep the deletes done during previously flushed segments |
CBufferedIndexInput | Base implementation class for buffered IndexInput |
CBufferedIndexOutput | Base implementation class for buffered IndexOutput |
CBufferedReader | Read text from a character-input stream, buffering characters so as to provide for the efficient reading of characters, arrays, and lines |
CByteBlockAllocator | |
CByteBlockPool | Class that Posting and PostingVector use to write byte streams into shared fixed-size byte[] arrays. The idea is to allocate slices of increasing lengths. For example, the first slice is 5 bytes, the next slice is 14, etc. We start by writing our bytes into the first 5 bytes. When we hit the end of the slice, we allocate the next slice and then write the address of the new slice into the last 4 bytes of the previous slice (the "forwarding address") |
CByteBlockPoolAllocatorBase | |
CByteCache | |
CByteComparator | Parses field's values as byte (using FieldCache#getBytes and sorts by ascending value |
CByteFieldSource | Obtains byte field values from the FieldCache using getBytes() and makes those values available as other numeric types, casting as needed |
CByteParser | Interface to parse bytes from document fields |
CByteSliceReader | IndexInput that knows how to read the byte slices written by Posting and PostingVector. We read the bytes in each slice until we hit the end of that slice at which point we read the forwarding address of the next slice and then jump to it |
CByteSliceWriter | Class to write byte streams into slices of shared byte[]. This is used by DocumentsWriter to hold the posting list for many terms in RAM |
CCache | Internal cache |
CCachingSpanFilter | Wraps another SpanFilter's result and caches it. The purpose is to allow filters to simply filter, and then wrap with this class to add caching |
CCachingTokenFilter | This class can be used if the token attributes of a TokenStream are intended to be consumed more than once. It caches all token attribute states locally in a List |
CCachingWrapperFilter | Wraps another filter's result and caches it. The purpose is to allow filters to simply filter, and then wrap with this class to add caching |
CCharArraySet | A simple class that stores Strings as char[]'s in a hash table. Note that this is not a general purpose class. For example, it cannot remove items from the set, nor does it resize its hash table to be smaller, etc. It is designed to be quick to test if a char[] is in the set without the necessity of converting it to a String first |
CCharBlockPool | |
CCharFilter | Subclasses of CharFilter can be chained to filter CharStream. They can be used as Reader with additional offset correction. Tokenizer s will automatically use correctOffset if a CharFilter/CharStream subclass is used |
CCharFolder | Utility class for folding character case |
CCharReader | CharReader is a Reader wrapper. It reads chars from Reader and outputs CharStream , defining an identify function correctOffset method that simply returns the provided offset |
CCharStream | CharStream adds correctOffset functionality over Reader . All Tokenizers accept a CharStream instead of Reader as input, which enables arbitrary character based filtering before tokenization. The correctOffset method fixed offsets to account for removal or insertion of characters, so that the offsets reported in the tokens match the character offsets of the original Reader |
CCharTokenizer | An abstract base class for simple, character-oriented tokenizers |
CCheckAbort | |
CCheckAbortNull | |
CCheckIndex | Basic tool and API to check the health of an index and write a new segments file that removes reference to problematic segments |
CChecksumIndexInput | Writes bytes through to a primary IndexInput, computing checksum as it goes. Note that you cannot use seek() |
CChecksumIndexOutput | Writes bytes through to a primary IndexOutput, computing checksum. Note that you cannot use seek() |
CCloseableThreadLocal | General purpose thread-local map |
CCollator | Convenience class for storing collate objects |
CCollection | Utility template class to handle collections that can be safely copied and shared |
CCollector | Collectors are primarily meant to be used to gather raw results from a search, and implement sorting or custom result filtering, collation, etc |
CCommitPoint | Holds details for each commit point. This class is also passed to the deletion policy. Note: this class has a natural ordering that is inconsistent with equals |
CComplexExplanation | Describes the score computation for document and query, and can distinguish a match independent of a positive value |
►CCompoundFileReader | Class for accessing a compound stream. This class implements a directory, but is limited to only read operations. Directory methods that would normally modify data throw an exception |
CFileEntry | |
►CCompoundFileWriter | Combines multiple files into a single compound file. The file format: VInt fileCount {Directory} fileCount entries with the following structure: int64_t dataOffset String fileName {File Data} fileCount entries with the raw data of the corresponding file |
CFileEntry | |
CCompressionTools | Simple utility class providing static methods to compress and decompress binary data for stored fields |
CConcurrentMergeScheduler | A MergeScheduler that runs each merge using a separate thread, up until a maximum number of threads (setMaxThreadCount ) at which when a merge is needed, the thread(s) that are updating the index will pause until one or more merges completes. This is a simple way to use concurrency in the indexing process without having to create and manage application level threads |
CConjunctionScorer | Scorer for conjunctions, sets of queries, all of which are required |
CConstants | Some useful Lucene constants |
CConstantScoreAutoRewrite | A rewrite method that tries to pick the best constant-score rewrite method based on term and document counts from the query. If both the number of terms and documents is small enough, then CONSTANT_SCORE_BOOLEAN_QUERY_REWRITE is used. Otherwise, CONSTANT_SCORE_FILTER_REWRITE is used |
CConstantScoreQuery | A query that wraps a filter and simply returns a constant score equal to the query boost for every document in the filter |
CCoordinator | |
CCountingConjunctionSumScorer | |
CCountingDisjunctionSumScorer | |
CCreationPlaceholder | |
CCSIndexInput | Implementation of an IndexInput that reads from a portion of the compound file |
CCustomScoreProvider | An instance of this subclass should be returned by CustomScoreQuery#getCustomScoreProvider , if you want to modify the custom score calculation of a CustomScoreQuery |
CCustomScoreQuery | Query that sets document score as a programmatic function of several (sub) scores: |
CCycleCheck | Debug utility to track shared_ptr utilization |
CCycleCheckT | |
CDateField | Provides support for converting dates to strings and vice-versa. The strings are structured so that lexicographic sorting orders by date, which makes them suitable for use as field values and search terms |
CDateTools | Provides support for converting dates to strings and vice-versa. The strings are structured so that lexicographic sorting orders them by date, which makes them suitable for use as field values and search terms |
CDefaultAttributeFactory | |
CDefaultIndexingChain | This is the current indexing chain: DocConsumer / DocConsumerPerThread --> code: DocFieldProcessor / DocFieldProcessorPerThread --> DocFieldConsumer / DocFieldConsumerPerThread / DocFieldConsumerPerField --> code: DocFieldConsumers / DocFieldConsumersPerThread / DocFieldConsumersPerField --> code: DocInverter / DocInverterPerThread / DocInverterPerField --> InvertedDocConsumer / InvertedDocConsumerPerThread / InvertedDocConsumerPerField --> code: TermsHash / TermsHashPerThread / TermsHashPerField --> TermsHashConsumer / TermsHashConsumerPerThread / TermsHashConsumerPerField --> code: FreqProxTermsWriter / FreqProxTermsWriterPerThread / FreqProxTermsWriterPerField --> code: TermVectorsTermsWriter / TermVectorsTermsWriterPerThread / TermVectorsTermsWriterPerField --> InvertedDocEndConsumer / InvertedDocConsumerPerThread / InvertedDocConsumerPerField --> code: NormsWriter / NormsWriterPerThread / NormsWriterPerField --> code: StoredFieldsWriter / StoredFieldsWriterPerThread / StoredFieldsWriterPerField |
CDefaultSimilarity | Default scoring implementation |
CDefaultSkipListReader | Implements the skip list reader for the default posting list format that stores positions and payloads |
CDefaultSkipListWriter | Implements the skip list writer for the default posting list format that stores positions and payloads |
CDirectory | A Directory is a flat list of files. Files may be written once, when they are created. Once a file is created it may only be opened for read, or deleted. Random access is permitted both when reading and writing. Directory locking is implemented by an instance of LockFactory , and can be changed for each Directory instance using setLockFactory |
CDirectoryReader | An IndexReader which reads indexes with multiple segments |
CDisjunctionMaxQuery | A query that generates the union of documents produced by its subqueries, and that scores each document with the maximum score for that document as produced by any subquery, plus a tie breaking increment for any additional matching subqueries. This is useful when searching for a word in multiple fields with different boost factors (so that the fields cannot be combined equivalently into a single search field). We want the primary score to be the one associated with the highest boost, not the sum of the field scores (as BooleanQuery would give). If the query is "albino elephant" this ensures that "albino" matching one field and "elephant" matching another gets a higher score than "albino" matching both fields. To get this result, use both BooleanQuery and DisjunctionMaxQuery: for each term a DisjunctionMaxQuery searches for it in each field, while the set of these DisjunctionMaxQuery's is combined into a BooleanQuery. The tie breaker capability allows results that include the same term in multiple fields to be judged better than results that include this term in only the best of those multiple fields, without confusing this with the better case of two different terms in the multiple fields |
CDisjunctionMaxScorer | The Scorer for DisjunctionMaxQuery. The union of all documents generated by the the subquery scorers is generated in document number order. The score for each document is the maximum of the scores computed by the subquery scorers that generate that document, plus tieBreakerMultiplier times the sum of the scores for the other subqueries that generate the document |
CDisjunctionSumScorer | A Scorer for OR like queries, counterpart of ConjunctionScorer. This Scorer implements Scorer#skipTo(int32_t) and uses skipTo() on the given Scorers |
CDocComparator | Sorts by ascending docID |
CDocConsumer | |
CDocConsumerPerThread | |
CDocFieldConsumer | |
CDocFieldConsumerPerField | |
CDocFieldConsumerPerThread | |
CDocFieldConsumers | This is just a "splitter" class: it lets you wrap two DocFieldConsumer instances as a single consumer |
CDocFieldConsumersPerDoc | |
CDocFieldConsumersPerField | |
CDocFieldConsumersPerThread | |
CDocFieldProcessor | This is a DocConsumer that gathers all fields under the same name, and calls per-field consumers to process field by field. This class doesn't doesn't do any "real" work of its own: it just forwards the fields to a DocFieldConsumer |
CDocFieldProcessorPerField | Holds all per thread, per field state |
CDocFieldProcessorPerThread | Gathers all Fieldables for a document under the same name, updates FieldInfos, and calls per-field consumers to process field by field |
CDocFieldProcessorPerThreadPerDoc | |
CDocIdBitSet | Simple DocIdSet and DocIdSetIterator backed by a BitSet |
CDocIdSet | A DocIdSet contains a set of doc ids. Implementing classes must only implement iterator to provide access to the set |
CDocIdSetIterator | This abstract class defines methods to iterate over a set of non-decreasing doc ids. Note that this class assumes it iterates on doc Ids, and therefore NO_MORE_DOCS is set to {@value NO_MORE_DOCS} in order to be used as a sentinel object. Implementations of this class are expected to consider INT_MAX as an invalid value |
CDocInverter | This is a DocFieldConsumer that inverts each field, separately, from a Document, and accepts a InvertedTermsConsumer to process those terms |
CDocInverterPerField | Holds state for inverting all occurrences of a single field in the document. This class doesn't do anything itself; instead, it forwards the tokens produced by analysis to its own consumer (InvertedDocConsumerPerField). It also interacts with an endConsumer (InvertedDocEndConsumerPerField) |
CDocInverterPerThread | This is a DocFieldConsumer that inverts each field, separately, from a Document, and accepts a InvertedTermsConsumer to process those terms |
CDocState | |
CDocument | Documents are the unit of indexing and search |
CDocumentsWriter | This class accepts multiple added documents and directly writes a single segment file. It does this more efficiently than creating a single segment per document (with DocumentWriter) and doing standard merges on those segments |
CDocumentsWriterThreadState | Used by DocumentsWriter to maintain per-thread state. We keep a separate Posting hash and other state for each thread and then merge postings hashes from all threads when writing the segment |
CDocValues | Represents field values as different types. Normally created via a ValueSuorce for a particular field and reader |
CDocWriter | Consumer returns this on each doc. This holds any state that must be flushed synchronized "in docID order". We gather these and flush them in order |
CDoubleCache | |
CDoubleComparator | Parses field's values as double (using FieldCache#getDoubles and sorts by ascending value |
CDoubleDocValues | |
CDoubleFieldSource | Obtains double field values from the FieldCache using getDoubles() and makes those values available as other numeric types, casting as needed |
CDoubleParser | Interface to parse doubles from document fields |
CEntry | |
CExactPhraseScorer | |
CExceptionTemplate | |
CExplanation | Describes the score computation for document and query |
CFastCharStream | An efficient implementation of QueryParserCharStream interface |
CField | |
CFieldable | Synonymous with Field |
CFieldCache | Maintains caches of term values |
CFieldCacheEntry | A unique Identifier/Description for each item in the FieldCache. Can be useful for logging/debugging |
CFieldCacheEntryImpl | |
CFieldCacheImpl | The default cache implementation, storing all values in memory. A WeakHashMap is used for storage |
CFieldCacheRangeFilter | A range filter built on top of a cached single term field (in FieldCache ) |
CFieldCacheSanityChecker | Provides methods for sanity checking that entries in the FieldCache are not wasteful or inconsistent |
CFieldCacheSource | A base class for ValueSource implementations that retrieve values for a single field from the FieldCache |
CFieldCacheTermsFilter | A Filter that only accepts documents whose single term value in the specified field is contained in the provided set of allowed terms |
CFieldComparator | A FieldComparator compares hits so as to determine their sort order when collecting the top results with TopFieldCollector . The concrete public FieldComparator classes here correspond to the SortField types |
CFieldComparatorSource | Provides a FieldComparator for custom field sorting |
CFieldDoc | A ScoreDoc which also contains information about how to sort the referenced document. In addition to the document number and score, this object contains an array of values for the document from the field(s) used to sort. For example, if the sort criteria was to sort by fields "a", "b" then "c", the fields object array will have three elements, corresponding respectively to the term values for the document in fields "a", "b" and "c". The class of each element in the array will be either Integer, Double or String depending on the type of values in the terms of each field |
CFieldDocSortedHitQueue | Collects sorted results from Searchable's and collates them. The elements put into this queue must be of type FieldDoc |
CFieldInfo | |
CFieldInfos | Access to the Fieldable Info file that describes document fields and whether or not they are indexed. Each segment has a separate Fieldable Info file. Objects of this class are thread-safe for multiple readers, but only one thread can be adding documents at a time, with no other reader or writer threads accessing this object |
CFieldInvertState | This class tracks the number and position / offset parameters of terms being added to the index. The information collected in this class is also used to calculate the normalization factor for a field |
CFieldMaskingSpanQuery | Wrapper to allow SpanQuery objects participate in composite single-field SpanQueries by 'lying' about their search field. That is, the masked SpanQuery will function as normal, but SpanQuery#getField() simply hands back the value supplied in this class's constructor |
CFieldNormStatus | Status from testing field norms |
CFieldScoreQuery | A query that scores each document as the value of the numeric input field |
CFieldSelector | The FieldSelector allows one to make decisions about what Fields get loaded on a Document by IndexReader#document(int32_t, FieldSelector) |
CFieldSortedTermVectorMapper | For each Field, store a sorted collection of TermVectorEntry s This is not thread-safe |
CFieldsReader | Class responsible for access to stored document fields. It uses <segment>.fdt and <segment>.fdx; files |
CFieldsWriter | |
CFieldValueHitQueue | A hit queue for sorting by hits by terms in more than one field. Uses FieldCache::DEFAULT for maintaining internal term lookup tables |
CFieldValueHitQueueEntry | |
CFileReader | Convenience class for reading character files |
CFileSwitchDirectory | A Directory instance that switches files between two other Directory instances |
CFilter | Abstract base class for restricting which documents may be returned during searching |
CFilteredDocIdSet | Abstract decorator class for a DocIdSet implementation that provides on-demand filtering/validation mechanism on a given DocIdSet |
CFilteredDocIdSetIterator | Abstract decorator class of a DocIdSetIterator implementation that provides on-demand filter/validation mechanism on an underlying DocIdSetIterator. See FilteredDocIdSet |
CFilteredQuery | A query that applies a filter to the results of another query |
CFilteredTermEnum | Abstract class for enumerating a subset of all terms |
CFilterIndexReader | A FilterIndexReader contains another IndexReader, which it uses as its basic source of data, possibly transforming the data along the way or providing additional functionality. The class FilterIndexReader itself simply implements all abstract methods of IndexReader with versions that pass all requests to the contained index reader. Subclasses of FilterIndexReader may further override some of these methods and may also provide additional methods and fields |
CFilterManager | Filter caching singleton. It can be used to save filters locally for reuse. Also could be used as a persistent storage for any filter as long as the filter provides a proper hashCode(), as that is used as the key in the cache |
CFilterTermDocs | Base class for filtering TermDocs implementations |
CFilterTermEnum | Base class for filtering TermEnum implementations |
CFilterTermPositions | Base class for filtering TermPositions implementations |
CFlagsAttribute | This attribute can be used to pass different flags down the tokenizer chain, eg from one TokenFilter to another one |
CFormatPostingsDocsConsumer | |
CFormatPostingsDocsWriter | Consumes doc & freq, writing them using the current index file format |
CFormatPostingsFieldsConsumer | Abstract API that consumes terms, doc, freq, prox and payloads postings. Concrete implementations of this actually do "something" with the postings (write it into the index in a specific format) |
CFormatPostingsFieldsWriter | |
CFormatPostingsPositionsConsumer | |
CFormatPostingsPositionsWriter | |
CFormatPostingsTermsConsumer | |
CFormatPostingsTermsWriter | |
CFreqProxFieldMergeState | Used by DocumentsWriter to merge the postings from multiple ThreadStates when creating a segment |
CFreqProxTermsWriter | |
CFreqProxTermsWriterPerField | |
CFreqProxTermsWriterPerThread | |
CFreqProxTermsWriterPostingList | |
CFSDirectory | Base class for Directory implementations that store index files in the file system. There are currently three core subclasses: |
CFSLockFactory | Base class for file system based locking implementation |
CFuture | A Future represents the result of an asynchronous computation. Methods are provided to check if the computation is complete, to wait for its completion, and to retrieve the result of the computation. The result can only be retrieved using method get when the computation has completed, blocking if necessary until it is ready |
CFuzzyQuery | Implements the fuzzy search query. The similarity measurement is based on the Levenshtein (edit distance) algorithm |
CFuzzyTermEnum | Subclass of FilteredTermEnum for enumerating all terms that are similar to the specified filter term |
CHashMap | Utility template class to handle hash maps that can be safely copied and shared |
CHashSet | Utility template class to handle hash set collections that can be safely copied and shared |
CHitQueue | |
CHitQueueBase | |
CIDFExplanation | Small Util class used to pass both an idf factor as well as an explanation for that factor |
CIndexCommit | Represents a single commit into an index as seen by the IndexDeletionPolicy or IndexReader |
CIndexDeletionPolicy | Policy for deletion of stale index commits . Implement this interface, and pass it to one of the IndexWriter or IndexReader constructors, to customize when older point-in-time commits are deleted from the index directory. The default deletion policy is KeepOnlyLastCommitDeletionPolicy , which always removes old commits as soon as a new commit is done (this matches the behavior before 2.2) |
CIndexFileDeleter | This class keeps track of each SegmentInfos instance that is still "live", either because it corresponds to a segments_N file in the Directory (a "commit", ie. a committed SegmentInfos) or because it's an in-memory SegmentInfos that a writer is actively updating but has not yet committed. This class uses simple reference counting to map the live SegmentInfos instances to individual files in the Directory |
CIndexFileNameFilter | Filename filter that accept filenames and extensions only created by Lucene |
CIndexFileNames | Constants representing filenames and extensions used by Lucene |
CIndexingChain | The IndexingChain must define the getChain(DocumentsWriter) method which returns the DocConsumer that the DocumentsWriter calls to process the documents |
CIndexInput | Abstract base class for input from a file in a Directory . A random-access input stream. Used for all Lucene index input operations |
CIndexOutput | Abstract base class for output to a file in a Directory. A random-access output stream. Used for all Lucene index output operations |
CIndexReader | IndexReader is an abstract class, providing an interface for accessing an index. Search of an index is done entirely through this abstract interface, so that any subclass which implements it is searchable |
CIndexReaderWarmer | If getReader has been called (ie, this writer is in near real-time mode), then after a merge completes, this class can be invoked to warm the reader on the newly merged segment, before the merge commits. This is not required for near real-time search, but will reduce search latency on opening a new near real-time reader after a merge completes |
CIndexSearcher | Implements search over a single IndexReader |
CIndexStatus | Returned from checkIndex() detailing the health and status of the index |
CIndexWriter | An IndexWriter creates and maintains an index |
CInfoStream | Utility class to support streaming info messages |
CInfoStreamFile | Stream override to write messages to a file |
CInfoStreamNull | Null stream override to eat messages |
CInfoStreamOut | Stream override to write messages to a std::cout |
CInputStreamReader | An InputStreamReader is a bridge from byte streams to character streams |
CInsanity | Simple container for a collection of related CacheEntry objects that in conjunction with each other represent some "insane" usage of the FieldCache |
CIntBlockPool | |
CIntCache | |
CIntComparator | Parses field's values as int (using FieldCache#getInts and sorts by ascending value |
CIntFieldSource | Obtains int field values from the FieldCache using getInts() and makes those values available as other numeric types, casting as needed |
CIntParser | Interface to parse ints from document fields |
CIntRangeBuilder | |
CInvertedDocConsumer | |
CInvertedDocConsumerPerField | |
CInvertedDocConsumerPerThread | |
CInvertedDocEndConsumer | |
CInvertedDocEndConsumerPerField | |
CInvertedDocEndConsumerPerThread | |
CISOLatin1AccentFilter | A filter that replaces accented characters in the ISO Latin 1 character set (ISO-8859-1) by their unaccented equivalent. The case will not be altered |
CKeepOnlyLastCommitDeletionPolicy | This IndexDeletionPolicy implementation that keeps only the most recent commit and immediately removes all prior commits after a new commit is done. This is the default deletion policy |
CKeywordAnalyzer | Tokenizes the entire stream as a single token. This is useful for data like zip codes, ids, and some product names |
CKeywordTokenizer | Emits the entire input as a single token |
CLazyField | |
CLengthFilter | Removes words that are too long or too short from the stream |
CLetterTokenizer | A LetterTokenizer is a tokenizer that divides text at non-letters. That's to say, it defines tokens as maximal strings of adjacent letters, as defined UnicodeUtil::isAlpha(c) predicate |
CLoadFirstFieldSelector | Load the First field and break. See FieldSelectorResult#LOAD_AND_BREAK |
CLock | An interprocess mutex lock |
CLockFactory | Base class for Locking implementation. Directory uses instances of this class to implement locking. Note that there are some useful tools to verify that your LockFactory is working correctly: VerifyingLockFactory , LockStressTest , LockVerifyServer |
CLogByteSizeMergePolicy | This is a LogMergePolicy that measures size of a segment as the total byte size of the segment's files |
CLogDocMergePolicy | This is a LogMergePolicy that measures size of a segment as the number of documents (not taking deletions into account) |
CLogMergePolicy | This class implements a MergePolicy that tries to merge segments into levels of exponentially increasing size, where each level has fewer segments than the value of the merge factor. Whenever extra segments (beyond the merge factor upper bound) are encountered, all segments within the level are merged. You can get or set the merge factor using getMergeFactor() and setMergeFactor(int) respectively |
CLongCache | |
CLongComparator | Parses field's values as long (using FieldCache#getLongs and sorts by ascending value |
CLongParser | Interface to parse longs from document fields |
CLongRangeBuilder | Callback for splitLongRange . You need to overwrite only one of the methods. NOTE: This is a very low-level interface, the method signatures may change in later versions |
CLowerCaseFilter | Normalizes token text to lower case |
CLowerCaseTokenizer | LowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together. It divides text at non-letters and converts them to lower case. While it is functionally equivalent to the combination of LetterTokenizer and LowerCaseFilter, there is a performance advantage to doing the two tasks at once, hence this (redundant) implementation |
CLuceneException | Lucene exception container |
CLuceneObject | Base class for all Lucene classes |
CLuceneSignal | Utility class to support signaling notifications |
CLuceneSync | Base class for all Lucene synchronised classes |
CLuceneThread | Lucene thread container |
CLuceneVersion | Use by certain classes to match version compatibility across releases of Lucene |
CMap | Utility template class to handle maps that can be safely copied and shared |
CMapFieldSelector | A FieldSelector based on a Map of field names to FieldSelectorResult s |
CMapOfSets | Helper class for keeping Lists of Objects associated with keys |
CMappingCharFilter | Simplistic CharFilter that applies the mappings contained in a NormalizeCharMap to the character stream, and correcting the resulting changes to the offsets |
CMatchAllDocsQuery | A query that matches all documents |
CMaxPayloadFunction | Returns the maximum payload score seen, else 1 if there are no payloads on the doc |
CMergeDocIDRemapper | Remaps docIDs after a merge has completed, where the merged segments had at least one deletion. This is used to renumber the buffered deletes in IndexWriter when a merge of segments with deletions commits |
CMergePolicy | A MergePolicy determines the sequence of primitive merge operations to be used for overall merge and optimize operations |
CMergeScheduler | IndexWriter uses an instance implementing this interface to execute the merges selected by a MergePolicy . The default MergeScheduler is ConcurrentMergeScheduler |
CMergeSpecification | A MergeSpecification instance provides the information necessary to perform multiple merges. It simply contains a list of OneMerge instances |
CMinPayloadFunction | Calculates the minimum payload seen |
CMiscUtils | |
CMMapDirectory | File-based Directory implementation that uses mmap for reading, and SimpleFSIndexOutput for writing |
CMultiFieldQueryParser | A QueryParser which constructs queries to search multiple fields |
CMultiLevelSkipListReader | This abstract class reads skip lists with multiple levels |
CMultiLevelSkipListWriter | This abstract class writes skip lists with multiple levels |
CMultiPhraseQuery | MultiPhraseQuery is a generalized version of PhraseQuery, with an added method add(Term[]) . To use this class, to search for the phrase "Microsoft app*" first use add(Term) on the term "Microsoft", then find all terms that have "app" as prefix using IndexReader.terms(Term), and use MultiPhraseQuery.add(Term[] terms) to add them to the query |
CMultipleTermPositions | Allows you to iterate over the TermPositions for multiple Term s as a single TermPositions |
CMultiReader | An IndexReader which reads multiple indexes, appending their content |
CMultiSearcher | Implements search over a set of Searchables |
CMultiTermDocs | |
CMultiTermEnum | |
CMultiTermPositions | |
CMultiTermQuery | An abstract Query that matches documents containing a subset of terms provided by a FilteredTermEnum enumeration |
CMultiTermQueryWrapperFilter | A wrapper for MultiTermQuery , that exposes its functionality as a Filter |
CNativeFSLockFactory | Implements LockFactory using native file lock |
CNearSpansOrdered | A Spans that is formed from the ordered subspans of a SpanNearQuery where the subspans do not overlap and have a maximum slop between them |
CNearSpansUnordered | Similar to NearSpansOrdered , but for the unordered case |
CNoLockFactory | Use this LockFactory to disable locking entirely. Only one instance of this lock is created. You should call getNoLockFactory() to get the instance |
CNormalizeCharMap | Holds a map of String input to String output, to be used with MappingCharFilter |
CNormsWriter | Writes norms. Each thread X field accumulates the norms for the doc/fields it saw, then the flush method below merges all of these together into a single _X.nrm file |
CNormsWriterPerField | Taps into DocInverter, as an InvertedDocEndConsumer, which is called at the end of inverting each field. We just look at the length for the field (docState.length) and record the norm |
CNormsWriterPerThread | |
CNum | Number of documents a delete term applies to |
CNumberTools | Provides support for converting longs to Strings, and back again. The strings are structured so that lexicographic sorting order is preserved |
CNumericComparator | |
CNumericField | This class provides a Field that enables indexing of numeric values for efficient range filtering and sorting. The native types int32_t, int64_t and double are directly supported. However, any value that can be converted into these native types can also be indexed. For example, date/time values represented by a Date can be translated into a int64_t value. If you don't need millisecond precision, you can quantize the value, either by dividing the result or using the separate getters (for year, month, etc.) to construct an int32_t or int64_t value |
CNumericRangeFilter | A Filter that only accepts numeric values within a specified range. To use this, you must first index the numeric values using NumericField (NumericTokenStream ) |
CNumericRangeQuery | A Query that matches numeric values within a specified range. To use this, you must first index the numeric values using NumericField (expert: NumericTokenStream ). If your terms are instead textual, you should use TermRangeQuery . NumericRangeFilter is the filter equivalent of this query |
CNumericTokenStream | This class provides a TokenStream for indexing numeric values that can be used by NumericRangeQuery or NumericRangeFilter |
CNumericUtils | This is a helper class to generate prefix-encoded representations for numerical values and supplies converters to represent double values as sortable integers/longs |
COffsetAttribute | The start and end character offset of a Token |
COneMerge | OneMerge provides the information necessary to perform an individual primitive merge operation, resulting in a single new segment. The merge spec includes the subset of segments to be merged as well as whether the new segment should use the compound file format |
COpenBitSet | An "open" BitSet implementation that allows direct access to the array of words storing the bits |
COpenBitSetDISI | |
COpenBitSetIterator | An iterator to iterate over set bits in an OpenBitSet. This is faster than nextSetBit() for iterating over the complete set of bits, especially when the density of the bits set is high |
COrdFieldSource | Obtains the ordinal of the field value from the default Lucene FieldCache using getStringIndex() |
CParallelArrayTermVectorMapper | Models the existing parallel array structure |
CParallelMultiSearcher | Implements parallel search over a set of Searchables |
CParallelReader | An IndexReader which reads multiple, parallel indexes. Each index added must have the same number of documents, but typically each contains different fields. Each document contains the union of the fields of all documents with the same document number. When searching, matches for a query term are from the first index added that has the field |
CParser | Marker interface as super-interface to all parsers. It is used to specify a custom parser to SortField#SortField(String, Parser) |
CPayload | A Payload is metadata that can be stored together with each occurrence of a term. This metadata is stored inline in the posting list of the specific term |
CPayloadAttribute | The start and end character offset of a Token |
CPayloadFunction | An abstract class that defines a way for Payload*Query instances to transform the cumulative effects of payload scores for a document |
CPayloadNearQuery | This class is very similar to SpanNearQuery except that it factors in the value of the payloads located at each of the positions where the TermSpans occurs |
CPayloadNearSpanScorer | |
CPayloadNearSpanWeight | |
CPayloadSpanUtil | Experimental class to get set of payloads for most standard Lucene queries. Operates like Highlighter - IndexReader should only contain doc of interest, best to use MemoryIndex |
CPayloadTermQuery | This class is very similar to SpanTermQuery except that it factors in the value of the payload located at each of the positions where the Term occurs |
CPerDocBuffer | RAMFile buffer for DocWriters |
CPerFieldAnalyzerWrapper | This analyzer is used to facilitate scenarios where different fields require different analysis techniques. Use addAnalyzer to add a non-default analyzer on a field name basis |
CPhrasePositions | Position of a term in a document that takes into account the term offset within the phrase |
CPhraseQuery | A Query that matches documents containing a particular sequence of terms. A PhraseQuery is built by QueryParser for input like "new york" |
CPhraseQueue | |
CPhraseScorer | Scoring functionality for phrase queries. A document is considered matching if it contains the phrase-query terms at "valid" positions. What "valid positions" are depends on the type of the phrase query: for an exact phrase query terms are required to appear in adjacent locations, while for a sloppy phrase query some distance between the terms is allowed. The abstract method phraseFreq() of extending classes is invoked for each document containing all the phrase query terms, in order to compute the frequency of the phrase query in that document. A non zero frequency means a match |
CPorterStemFilter | Transforms the token stream as per the Porter stemming algorithm. Note: the input to the stemming filter must already be in lower case, so you will need to use LowerCaseFilter or LowerCaseTokenizer further down the Tokenizer chain in order for this to work properly |
CPorterStemmer | This is the Porter stemming algorithm, coded up as thread-safe ANSI C by the author |
CPositionBasedTermVectorMapper | |
CPositionIncrementAttribute | The positionIncrement determines the position of this token relative to the previous Token in a TokenStream, used in phrase searching |
CPositionInfo | |
CPositiveScoresOnlyCollector | A Collector implementation which wraps another Collector and makes sure only documents with scores > 0 are collected |
CPrefixFilter | A Filter that restricts search results to values that have a matching prefix in a given field |
CPrefixQuery | A Query that matches documents containing terms with a specified prefix. A PrefixQuery is built by QueryParser for input like app* |
CPrefixTermEnum | Subclass of FilteredTermEnum for enumerating all terms that match the specified prefix filter term |
CPriorityQueue | A PriorityQueue maintains a partial ordering of its elements such that the least element can always be found in constant time. Put()'s and pop()'s require log(size) time |
CPriorityQueueScoreDocs | |
CQuery | The abstract base class for queries |
CQueryParseError | Utility class to handle query parse errors |
►CQueryParser | The most important method is parse(const String&) |
CJJCalls | |
CQueryParserCharStream | This interface describes a character stream that maintains line and column number positions of the characters. It also has the capability to backup the stream to some extent. An implementation of this interface is used in the QueryParserTokenManager |
CQueryParserConstants | Token literal values and constants |
CQueryParserToken | Describes the input token stream |
CQueryParserTokenManager | Token Manager |
CQueryTermVector | |
CQueryWrapperFilter | Constrains search results to only match those which also match a provided query |
CRAMDirectory | A memory-resident Directory implementation. Locking implementation is by default the SingleInstanceLockFactory but can be changed with setLockFactory . Lock acquisition sequence: RAMDirectory, then RAMFile |
CRAMFile | File used as buffer in RAMDirectory |
CRAMInputStream | A memory-resident IndexInput implementation |
CRAMOutputStream | A memory-resident IndexOutput implementation |
CRandom | Utility class to generate a stream of pseudorandom numbers |
CRawPostingList | This is the base class for an in-memory posting list, keyed by a Token. TermsHash maintains a hash table holding one instance of this per unique Token. Consumers of TermsHash (TermsHashConsumer ) must subclass this class with its own concrete class. FreqProxTermsWriterPostingList is a private inner class used for the freq/prox postings, and TermVectorsTermsWriterPostingList is a private inner class used to hold TermVectors postings |
CReader | Abstract class for reading character streams |
CReaderCommit | |
CReaderUtil | Common util methods for dealing with IndexReader s |
CReadOnlyDirectoryReader | |
CReadOnlySegmentReader | |
CRefCount | Tracks the reference count for a single index file |
CRelevanceComparator | Sorts by descending relevance. NOTE: if you are sorting only by descending relevance and then secondarily by ascending docID, performance is faster using TopScoreDocCollector directly (which IndexSearcher#search uses when no Sort is specified) |
CReqExclScorer | A Scorer for queries with a required subscorer and an excluding (prohibited) sub DocIdSetIterator. This Scorer implements Scorer#skipTo(int32_t) , and it uses the skipTo() on the given scorers |
CReqOptSumScorer | A Scorer for queries with a required part and an optional part. Delays skipTo() on the optional part until a score() is needed. This Scorer implements Scorer#skipTo(int32_t) |
CReusableStringReader | Used by DocumentsWriter to implemented a StringReader that can be reset to a new string; we use this when tokenizing the string value from a Field |
CReverseOrdFieldSource | Obtains the ordinal of the field value from the default Lucene FieldCache using getStringIndex() and reverses the order |
CRewriteMethod | Abstract class that defines how the query is rewritten |
CScoreCachingWrappingScorer | A Scorer which wraps another scorer and caches the score of the current document. Successive calls to score() will return the same result and will not invoke the wrapped Scorer's score() method, unless the current document has changed |
CScoreDoc | Expert: Returned by low-level search implementations |
CScorer | Common scoring functionality for different types of queries |
CScorerDocQueue | A ScorerDocQueue maintains a partial ordering of its Scorers such that the least Scorer can always be found in constant time. Put()'s and pop()'s require log(size) time. The ordering is by Scorer::doc() |
CScorerVisitor | |
CSearchable | The interface for search implementations |
CSearcher | An abstract base class for search implementations. Implements the main search methods |
CSegmentInfo | Information about a segment such as it's name, directory, and files related to the segment |
CSegmentInfoCollection | A collection of SegmentInfo objects to be used as a base class for SegmentInfos |
CSegmentInfos | A collection of SegmentInfo objects with methods for operating on those segments in relation to the file system |
CSegmentInfoStatus | Holds the status of each segment in the index. See segmentInfos |
CSegmentMergeInfo | |
CSegmentMergeQueue | |
CSegmentMerger | Combines two or more Segments, represented by an IndexReader (add , into a single Segment. After adding the appropriate readers, call the merge method to combine the segments |
CSegmentReader | |
CSegmentTermDocs | |
CSegmentTermEnum | |
CSegmentTermPositions | |
CSegmentTermPositionVector | |
CSegmentTermVector | |
CSegmentWriteState | |
CSerialMergeScheduler | A MergeScheduler that simply does each merge sequentially, using the current thread |
CSet | Utility template class to handle set based collections that can be safely copied and shared |
CSetBasedFieldSelector | Declare what fields to load normally and what fields to load lazily |
CSimilarity | Scoring API |
CSimilarityDelegator | Delegating scoring implementation. Useful in Query#getSimilarity(Searcher) implementations, to override only certain methods of a Searcher's Similarity implementation |
CSimpleAnalyzer | An Analyzer that filters LetterTokenizer with LowerCaseFilter |
CSimpleFSDirectory | A straightforward implementation of FSDirectory using std::ofstream and std::ifstream |
CSimpleFSLockFactory | Implements LockFactory using File#createNewFile() |
CSimpleLRUCache | General purpose LRU cache map. Accessing an entry will keep the entry cached. get(const KEY&) and put(const KEY&, const VALUE&) results in an access to the corresponding entry |
CSingleInstanceLockFactory | Implements LockFactory for a single in-process instance, meaning all locking will take place through this one instance. Only use this LockFactory when you are certain all IndexReaders and IndexWriters for a given index are running against a single shared in-process Directory instance. This is currently the default locking for RAMDirectory |
CSingleMatchScorer | Count a scorer as a single match |
CSingleTermEnum | Subclass of FilteredTermEnum for enumerating a single term |
CSingleTokenAttributeSource | |
CSinkFilter | |
CSinkTokenStream | A filter that decides which AttributeSource states to store in the sink |
CSkipBuffer | Used to buffer the top skip levels |
CSkipDocWriter | |
CSloppyPhraseScorer | |
CSmallDouble | Floating point numbers smaller than 32 bits |
CSnapshotDeletionPolicy | |
CSort | Encapsulates sort criteria for returned hits |
CSortedTermVectorMapper | Store a sorted collection of TermVectorEntry s. Collects all term information into a single, sorted set |
CSortedVIntList | Stores and iterate on sorted integers in compressed form in RAM |
CSortField | Stores information about how to sort documents by terms in an individual field. Fields must be indexed in order to sort by them |
CSpanFilter | Abstract base class providing a mechanism to restrict searches to a subset of an index and also maintains and returns position information |
CSpanFilterResult | The results of a SpanQueryFilter. Wraps the BitSet and the position information from the SpanQuery |
CSpanFirstQuery | Matches spans near the beginning of a field |
CSpanNearQuery | Matches spans which are near one another. One can specify slop, the maximum number of intervening unmatched positions, as well as whether matches are required to be in-order |
CSpanNotQuery | Removes matches which overlap with another SpanQuery |
CSpanOrQuery | Matches the union of its clauses |
CSpanQuery | Base class for span-based queries |
CSpanQueryFilter | Constrains search results to only match those which also match a provided query. Also provides position information about where each document matches at the cost of extra space compared with the QueryWrapperFilter. There is an added cost to this above what is stored in a QueryWrapperFilter . Namely, the position information for each matching document is stored |
CSpans | An enumeration of span matches. Used to implement span searching. Each span represents a range of term positions within a document. Matches are enumerated in order, by increasing document number, within that by increasing start position and finally by increasing end position |
CSpanScorer | Public for extension only |
CSpanTermQuery | Matches spans containing a term |
CSpanWeight | Public for use by other weight implementations |
CStandardAnalyzer | Filters StandardTokenizer with StandardFilter , LowerCaseFilter and StopFilter , using a list of English stop words |
CStandardFilter | Normalizes tokens extracted with StandardTokenizer |
CStandardTokenizer | A grammar-based tokenizer |
CStandardTokenizerImpl | |
CStartEnd | |
CStopAnalyzer | Filters LetterTokenizer with LowerCaseFilter and StopFilter |
CStopFilter | Removes stop words from a token stream |
CStoredFieldStatus | Status from testing stored fields |
CStoredFieldsWriter | This is a DocFieldConsumer that writes stored fields |
CStoredFieldsWriterPerDoc | |
CStoredFieldsWriterPerThread | |
CStringCache | |
CStringComparatorLocale | Sorts by a field's value using the Collator for a given Locale |
CStringIndex | Stores term text values and document ordering data |
CStringIndexCache | |
CStringOrdValComparator | Sorts by field's natural String sort order, using ordinals. This is functionally equivalent to StringValComparator , but it first resolves the string to their relative ordinal positions (using the index returned by FieldCache#getStringIndex ), and does most comparisons using the ordinals. For medium to large results, this comparator will be much faster than StringValComparator . For very small result sets it may be slower |
CStringReader | Convenience class for reading strings |
CStringUtils | |
CStringValComparator | Sorts by field's natural String sort order. All comparisons are done using String.compare, which is slow for medium to large result sets but possibly very fast for very small results sets |
CSubScorer | |
CSynchronize | Utility class to support locking via a mutex |
CSyncLock | Utility class to support scope locking |
CTeeSinkTokenFilter | This TokenFilter provides the ability to set aside attribute states that have already been analyzed. This is useful in situations where multiple fields share many common analysis steps and then go their separate ways |
CTerm | A Term represents a word from text. This is the unit of search. It is composed of two elements, the text of the word, as a string, and the name of the field that the text occurred in, an interned string |
CTermAttribute | The term text of a Token |
CTermBuffer | |
CTermDocs | TermDocs provides an interface for enumerating <document, frequency>; pairs for a term. The document portion names each document containing the term. Documents are indicated by number. The frequency portion gives the number of times the term occurred in each document. The pairs are ordered by document number |
CTermEnum | Abstract class for enumerating terms |
CTermFreqVector | Provides access to stored term vector of a document field. The vector consists of the name of the field, an array of the terms that occur in the field of the Document and a parallel array of frequencies. Thus, getTermFrequencies()[5] corresponds with the frequency of getTerms()[5], assuming there are at least 5 terms in the Document |
CTermIndexStatus | Status from testing term index |
CTermInfo | A TermInfo is the record of information stored for a term |
CTermInfosReader | This stores a monotonically increasing set of <Term, TermInfo> pairs in a Directory. Pairs are accessed either by Term or by ordinal position the set |
CTermInfosReaderThreadResources | |
CTermInfosWriter | This stores a monotonically increasing set of <Term, TermInfo> pairs in a Directory. A TermInfos can be written once, in order |
CTermPositions | TermPositions provides an interface for enumerating the <document, frequency, <position>*> tuples for a term. The document and frequency are the same as for a TermDocs. The positions portion lists the ordinal positions of each occurrence of a term in a document |
CTermPositionVector | Extends TermFreqVector to provide additional information about positions in which each of the terms is found. A TermPositionVector not necessarily contains both positions and offsets, but at least one of these arrays exists |
CTermQuery | A Query that matches documents containing a term. This may be combined with other terms with a BooleanQuery |
CTermRangeFilter | A Filter that restricts search results to a range of term values in a given field |
CTermRangeQuery | A Query that matches documents within an range of terms |
CTermRangeTermEnum | Subclass of FilteredTermEnum for enumerating all terms that match the specified range parameters |
CTermScorer | A Scorer for documents matching a Term |
CTermsHash | This class implements InvertedDocConsumer , which is passed each token produced by the analyzer on each field. It stores these tokens in a hash table, and allocates separate byte streams per token. Consumers of this class, eg FreqProxTermsWriter and TermVectorsTermsWriter , write their own byte streams under each term |
CTermsHashConsumer | |
CTermsHashConsumerPerField | Implement this class to plug into the TermsHash processor, which inverts & stores Tokens into a hash table and provides an API for writing bytes into multiple streams for each unique Token |
CTermsHashConsumerPerThread | |
CTermsHashPerField | |
CTermsHashPerThread | |
CTermSpans | Public for extension only |
CTermVectorEntry | Convenience class for holding TermVector information |
CTermVectorEntryFreqSortedComparator | Compares TermVectorEntry s first by frequency and then by the term (case-sensitive) |
CTermVectorMapper | The TermVectorMapper can be used to map Term Vectors into your own structure instead of the parallel array structure used by IndexReader#getTermFreqVector(int,String) |
CTermVectorOffsetInfo | Holds information pertaining to a Term in a TermPositionVector 's offset information. This offset information is the character offset as set during the Analysis phase (and thus may not be the actual offset in the original content) |
CTermVectorsPositionInfo | Container for a term at a position |
CTermVectorsReader | |
CTermVectorStatus | Status from testing stored fields |
CTermVectorsTermsWriter | |
CTermVectorsTermsWriterPerDoc | |
CTermVectorsTermsWriterPerField | |
CTermVectorsTermsWriterPerThread | |
CTermVectorsTermsWriterPostingList | |
CTermVectorsWriter | |
CTestPoint | Used for unit testing as a substitute for stack trace |
CTestScope | |
CThreadPool | Utility class to handle a pool of threads |
CTimeLimitingCollector | The TimeLimitingCollector is used to timeout search requests that take longer than the maximum allowed search time limit. After this time is exceeded, the search thread is stopped by throwing a TimeExceededException |
CToken | A Token is an occurrence of a term from the text of a field. It consists of a term's text, the start and end offset of the term in the text of the field and a type string |
CTokenAttributeFactory | Creates a TokenAttributeFactory returning Token as instance for the basic attributes and for all other attributes calls the given delegate factory |
CTokenFilter | A TokenFilter is a TokenStream whose input is another TokenStream |
CTokenizer | A Tokenizer is a TokenStream whose input is a Reader |
CTokenStream | A TokenStream enumerates the sequence of tokens, either from Field s of a Document or from query text |
CTopDocs | Represents hits returned by Searcher#search(QueryPtr, FilterPtr, int32_t) and Searcher#search(QueryPtr, int32_t) |
CTopDocsCollector | A base class for all collectors that return a TopDocs output. This collector allows easy extension by providing a single constructor which accepts a PriorityQueue as well as protected members for that priority queue and a counter of the number of total hits |
CTopFieldCollector | A Collector that sorts by SortField using FieldComparator s |
CTopFieldDocs | Represents hits returned by Searcher#search(QueryPtr, FilterPtr, int32_t, SortPtr) |
CTopScoreDocCollector | A Collector implementation that collects the top-scoring hits, returning them as a TopDocs . This is used by IndexSearcher to implement TopDocs -based search. Hits are sorted by score descending and then (when the scores are tied) docID ascending. When you create an instance of this collector you should know in advance whether documents are going to be collected in doc Id order or not |
CTranslationResult | Utility class that contains utf8 and unicode translations |
CTypeAttribute | A Token's lexical type. The Default value is "word" |
CUnicodeResult | |
CUnicodeUtil | |
CUTF16Decoder | |
CUTF8Base | |
CUTF8Decoder | |
CUTF8DecoderStream | |
CUTF8Encoder | |
CUTF8EncoderStream | |
CUTF8Result | |
CValueSource | Source of values for basic function queries |
CValueSourceQuery | A Query that sets the scores of document to the values obtained from a ValueSource |
CVariantUtils | |
CWaitQueue | |
CWeakHashMap | Utility template class to handle weak keyed maps |
CWeight | Calculate query weights and build query scorers |
CWhitespaceAnalyzer | An Analyzer that uses WhitespaceTokenizer |
CWhitespaceTokenizer | A WhitespaceTokenizer is a tokenizer that divides text at whitespace. Adjacent sequences of non-Whitespace characters form tokens |
CWildcardQuery | Implements the wildcard search query. Supported wildcards are *, which matches any character sequence (including the empty one), and ?, which matches any single character. Note this query can be slow, as it needs to iterate over many terms. In order to prevent extremely slow WildcardQueries, a Wildcard term should not start with one of the wildcards * or ? |
CWildcardTermEnum | Subclass of FilteredTermEnum for enumerating all terms that match the specified wildcard filter term |
CWordlistLoader | Loader for text files that represent a list of stopwords |