Dependencies
Declarations
-
org
-
apache
-
lucene
-
search
-
highlight
-
DefaultEncoder
-
Encoder
-
Formatter
-
Fragmenter
-
GradientFormatter
-
Highlighter
-
DEFAULT_MAX_CHARS_TO_ANALYZE: int
-
formatter: Formatter
-
encoder: Encoder
-
fragmentScorer: Scorer
-
maxDocCharsToAnalyze: int
-
textFragmenter: Fragmenter
-
Highlighter(Scorer): void
-
Highlighter(Formatter, Scorer): void
-
Highlighter(Formatter, Encoder, Scorer): void
-
getBestFragment(Analyzer, String, String): String
-
getBestFragment(TokenStream, String): String
-
getBestFragments(Analyzer, String, String, int): String[]
-
getBestFragments(TokenStream, String, int): String[]
-
getBestTextFragments(TokenStream, String, boolean, int): TextFragment[]
-
mergeContiguousFragments(TextFragment[]): void
-
getBestFragments(TokenStream, String, int, String): String
-
getMaxDocCharsToAnalyze(): int
-
setMaxDocCharsToAnalyze(int): void
-
getTextFragmenter(): Fragmenter
-
setTextFragmenter(Fragmenter): void
-
getFragmentScorer(): Scorer
-
setFragmentScorer(Scorer): void
-
getEncoder(): Encoder
-
setEncoder(Encoder): void
-
ensureArgumentNotNull(Object, String): void
-
FragmentQueue
-
InvalidTokenOffsetsException
-
LimitTokenOffsetFilter
-
NullFragmenter
-
OffsetLimitTokenFilter
-
PositionSpan
-
QueryScorer
-
getTokenScore(): float
-
totalScore: float
-
foundTerms: Set<String>
-
fieldWeightedSpanTerms: Map<String, WeightedSpanTerm>
-
maxTermWeight: float
-
position: int
-
defaultField: String
-
termAtt: CharTermAttribute
-
posIncAtt: PositionIncrementAttribute
-
expandMultiTermQuery: boolean
-
query: Query
-
field: String
-
reader: IndexReader
-
skipInitExtractor: boolean
-
wrapToCaching: boolean
-
maxCharsToAnalyze: int
-
usePayloads: boolean
-
QueryScorer(Query): void
-
QueryScorer(Query, String): void
-
QueryScorer(Query, IndexReader, String): void
-
QueryScorer(Query, IndexReader, String, String): void
-
QueryScorer(Query, String, String): void
-
QueryScorer(WeightedSpanTerm[]): void
-
getFragmentScore(): float
-
getMaxTermWeight(): float
-
init(TokenStream): TokenStream
-
getWeightedSpanTerm(String): WeightedSpanTerm
-
init(Query, String, IndexReader, boolean): void
-
initExtractor(TokenStream): TokenStream
-
newTermExtractor(String): WeightedSpanTermExtractor
-
startFragment(TextFragment): void
-
isExpandMultiTermQuery(): boolean
-
setExpandMultiTermQuery(boolean): void
-
isUsePayloads(): boolean
-
setUsePayloads(boolean): void
-
setWrapIfNotCachingTokenFilter(boolean): void
-
setMaxDocCharsToAnalyze(int): void
-
QueryTermExtractor
-
QueryTermScorer
-
Scorer
-
SimpleFragmenter
-
SimpleHTMLEncoder
-
SimpleHTMLFormatter
-
SimpleSpanFragmenter
-
SpanGradientFormatter
-
TermVectorLeafReader
-
TextFragment
-
TokenGroup
-
TokenSources
-
TokenSources(): void
-
getTokenStream(String, Fields, String, Analyzer, int): TokenStream
-
getTermVectorTokenStreamOrNull(String, Fields, int): TokenStream
-
getAnyTokenStream(IndexReader, int, String, Document, Analyzer): TokenStream
-
getAnyTokenStream(IndexReader, int, String, Analyzer): TokenStream
-
getTokenStream(Terms, boolean): TokenStream
-
getTokenStream(Terms): TokenStream
-
getTokenStreamWithOffsets(IndexReader, int, String): TokenStream
-
getTokenStream(IndexReader, int, String, Analyzer): TokenStream
-
getTokenStream(Document, String, Analyzer): TokenStream
-
getTokenStream(String, String, Analyzer): TokenStream
-
TokenStreamFromTermVector
-
WeightedSpanTerm
-
WeightedSpanTermExtractor
-
fieldName: String
-
tokenStream: TokenStream
-
defaultField: String
-
expandMultiTermQuery: boolean
-
cachedTokenStream: boolean
-
wrapToCaching: boolean
-
maxDocCharsToAnalyze: int
-
usePayloads: boolean
-
internalReader: LeafReader
-
WeightedSpanTermExtractor(): void
-
WeightedSpanTermExtractor(String): void
-
extract(Query, float, Map<String, WeightedSpanTerm>): void
-
isQueryUnsupported(Class<Query>): boolean
-
extractUnknownQuery(Query, Map<String, WeightedSpanTerm>): void
-
extractWeightedSpanTerms(Map<String, WeightedSpanTerm>, SpanQuery, float): void
-
extractWeightedTerms(Map<String, WeightedSpanTerm>, Query, float): void
-
fieldNameComparator(String): boolean
-
getLeafContext(): LeafReaderContext
-
DelegatingLeafReader
-
getWeightedSpanTerms(Query, float, TokenStream): Map<String, WeightedSpanTerm>
-
getWeightedSpanTerms(Query, float, TokenStream, String): Map<String, WeightedSpanTerm>
-
getWeightedSpanTermsWithScores(Query, float, TokenStream, String, IndexReader): Map<String, WeightedSpanTerm>
-
collectSpanQueryFields(SpanQuery, Set<String>): void
-
mustRewriteQuery(SpanQuery): boolean
-
PositionCheckingMap
-
getExpandMultiTermQuery(): boolean
-
setExpandMultiTermQuery(boolean): void
-
isUsePayloads(): boolean
-
setUsePayloads(boolean): void
-
isCachedTokenStream(): boolean
-
getTokenStream(): TokenStream
-
setWrapIfNotCachingTokenFilter(boolean): void
-
setMaxDocCharsToAnalyze(int): void
-
WeightedTerm
-
uhighlight
-
AnalysisOffsetStrategy
-
CustomSeparatorBreakIterator
-
DefaultPassageFormatter
-
FieldHighlighter
-
field: String
-
fieldOffsetStrategy: FieldOffsetStrategy
-
breakIterator: BreakIterator
-
passageScorer: PassageScorer
-
maxPassages: int
-
maxNoHighlightPassages: int
-
passageFormatter: PassageFormatter
-
FieldHighlighter(String, FieldOffsetStrategy, BreakIterator, PassageScorer, int, int, PassageFormatter): void
-
getField(): String
-
getOffsetSource(): OffsetSource
-
highlightFieldForDoc(LeafReader, int, String): Object
-
getSummaryPassagesNoHighlight(int): Passage[]
-
highlightOffsetsEnums(OffsetsEnum): Passage[]
-
maybeAddPassage(PriorityQueue<Passage>, PassageScorer, Passage, int): Passage
-
FieldOffsetStrategy
-
components: UHComponents
-
FieldOffsetStrategy(UHComponents): void
-
getField(): String
-
getOffsetSource(): OffsetSource
-
getOffsetsEnum(LeafReader, int, String): OffsetsEnum
-
createOffsetsEnumFromReader(LeafReader, int): OffsetsEnum
-
createOffsetsEnumsWeightMatcher(LeafReader, int, List<OffsetsEnum>): void
-
createOffsetsEnumsForTerms(BytesRef[], Terms, int, List<OffsetsEnum>): void
-
createOffsetsEnumsForAutomata(Terms, int, List<OffsetsEnum>): void
-
LengthGoalBreakIterator
-
MemoryIndexOffsetStrategy
-
MultiTermHighlighting
-
NoOpOffsetStrategy
-
OffsetsEnum
-
OverlaySingleDocTermsLeafReader
-
Passage
-
PassageFormatter
-
PassageScorer
-
PhraseHelper
-
PostingsOffsetStrategy
-
PostingsWithTermVectorsOffsetStrategy
-
SplittingBreakIterator
-
TermVectorFilteredLeafReader
-
TermVectorOffsetStrategy
-
TokenStreamOffsetStrategy
-
UHComponents
-
UnifiedHighlighter
-
MULTIVAL_SEP_CHAR: char
-
DEFAULT_MAX_LENGTH: int
-
DEFAULT_CACHE_CHARS_THRESHOLD: int
-
EMPTY_INDEXSEARCHER: IndexSearcher
-
static class initializer
-
ZERO_LEN_AUTOMATA_ARRAY: CharacterRunAutomaton[]
-
searcher: IndexSearcher
-
indexAnalyzer: Analyzer
-
defaultHandleMtq: boolean
-
defaultHighlightPhrasesStrictly: boolean
-
defaultPassageRelevancyOverSpeed: boolean
-
maxLength: int
-
defaultBreakIterator: Supplier<BreakIterator>
-
defaultFieldMatcher: Predicate<String>
-
defaultScorer: PassageScorer
-
defaultFormatter: PassageFormatter
-
defaultMaxNoHighlightPassages: int
-
fieldInfos: FieldInfos
-
cacheFieldValCharsThreshold: int
-
extractTerms(Query): Set<Term>
-
UnifiedHighlighter(IndexSearcher, Analyzer): void
-
setHandleMultiTermQuery(boolean): void
-
setHighlightPhrasesStrictly(boolean): void
-
setMaxLength(int): void
-
setBreakIterator(Supplier<BreakIterator>): void
-
setScorer(PassageScorer): void
-
setFormatter(PassageFormatter): void
-
setMaxNoHighlightPassages(int): void
-
setCacheFieldValCharsThreshold(int): void
-
setFieldMatcher(Predicate<String>): void
-
shouldHandleMultiTermQuery(String): boolean
-
shouldHighlightPhrasesStrictly(String): boolean
-
shouldPreferPassageRelevancyOverSpeed(String): boolean
-
getFieldMatcher(String): Predicate<String>
-
getMaxLength(): int
-
getBreakIterator(String): BreakIterator
-
getScorer(String): PassageScorer
-
getFormatter(String): PassageFormatter
-
getMaxNoHighlightPassages(String): int
-
getCacheFieldValCharsThreshold(): int
-
getIndexSearcher(): IndexSearcher
-
getIndexAnalyzer(): Analyzer
-
OffsetSource
-
getOffsetSource(String): OffsetSource
-
getFieldInfo(String): FieldInfo
-
highlight(String, Query, TopDocs): String[]
-
highlight(String, Query, TopDocs, int): String[]
-
highlightFields(String[], Query, TopDocs): Map<String, String[]>
-
highlightFields(String[], Query, TopDocs, int[]): Map<String, String[]>
-
highlightFields(String[], Query, int[], int[]): Map<String, String[]>
-
highlightFieldsAsObjects(String[], Query, int[], int[]): Map<String, Object[]>
-
calculateOptimalCacheCharsThreshold(int, int): int
-
copyAndSortFieldsWithMaxPassages(String[], int[], String[], int[]): void
-
copyAndSortDocIdsWithIndex(int[], int[], int[]): void
-
highlightWithoutSearcher(String, Query, String, int): Object
-
getFieldHighlighter(String, Query, Set<Term>, int): FieldHighlighter
-
getHighlightComponents(String, Query, Set<Term>): UHComponents
-
hasUnrecognizedQuery(Predicate<String>, Query): boolean
-
filterExtractedTerms(Predicate<String>, Set<Term>): BytesRef[]
-
getFlags(String): Set<HighlightFlag>
-
getPhraseHelper(String, Query, Set<HighlightFlag>): PhraseHelper
-
getAutomata(String, Query, Set<HighlightFlag>): CharacterRunAutomaton[]
-
getOptimizedOffsetSource(UHComponents): OffsetSource
-
getOffsetStrategy(OffsetSource, UHComponents): FieldOffsetStrategy
-
requiresRewrite(SpanQuery): Boolean
-
preSpanQueryRewrite(Query): Collection<Query>
-
asDocIdSetIterator(int[]): DocIdSetIterator
-
loadFieldValues(String[], DocIdSetIterator, int): List<CharSequence[]>
-
newLimitedStoredFieldsVisitor(String[]): LimitedStoredFieldVisitor
-
LimitedStoredFieldVisitor
-
TermVectorReusingLeafReader
-
HighlightFlag
-
WholeBreakIterator
-
vectorhighlight
-
BaseFragListBuilder
-
BaseFragmentsBuilder
-
preTags: String[]
-
postTags: String[]
-
COLORED_PRE_TAGS: String[]
-
COLORED_POST_TAGS: String[]
-
multiValuedSeparator: char
-
boundaryScanner: BoundaryScanner
-
discreteMultiValueHighlighting: boolean
-
BaseFragmentsBuilder(): void
-
BaseFragmentsBuilder(String[], String[]): void
-
BaseFragmentsBuilder(BoundaryScanner): void
-
BaseFragmentsBuilder(String[], String[], BoundaryScanner): void
-
checkTagsArgument(Object): Object
-
getWeightedFragInfoList(List<WeightedFragInfo>): List<WeightedFragInfo>
-
NULL_ENCODER: Encoder
-
createFragment(IndexReader, int, String, FieldFragList): String
-
createFragments(IndexReader, int, String, FieldFragList, int): String[]
-
createFragment(IndexReader, int, String, FieldFragList, String[], String[], Encoder): String
-
createFragments(IndexReader, int, String, FieldFragList, int, String[], String[], Encoder): String[]
-
getFields(IndexReader, int, String): Field[]
-
makeFragment(StringBuilder, int[], Field[], WeightedFragInfo, String[], String[], Encoder): String
-
getFragmentSourceMSO(StringBuilder, int[], Field[], int, int, int[]): String
-
getFragmentSource(StringBuilder, int[], Field[], int, int): String
-
discreteMultiValueHighlighting(List<WeightedFragInfo>, Field[]): List<WeightedFragInfo>
-
setMultiValuedSeparator(char): void
-
getMultiValuedSeparator(): char
-
isDiscreteMultiValueHighlighting(): boolean
-
setDiscreteMultiValueHighlighting(boolean): void
-
getPreTag(int): String
-
getPostTag(int): String
-
getPreTag(String[], int): String
-
getPostTag(String[], int): String
-
BoundaryScanner
-
BreakIteratorBoundaryScanner
-
FastVectorHighlighter
-
DEFAULT_PHRASE_HIGHLIGHT: boolean
-
DEFAULT_FIELD_MATCH: boolean
-
phraseHighlight: boolean
-
fieldMatch: boolean
-
fragListBuilder: FragListBuilder
-
fragmentsBuilder: FragmentsBuilder
-
phraseLimit: int
-
FastVectorHighlighter(): void
-
FastVectorHighlighter(boolean, boolean): void
-
FastVectorHighlighter(boolean, boolean, FragListBuilder, FragmentsBuilder): void
-
getFieldQuery(Query): FieldQuery
-
getFieldQuery(Query, IndexReader): FieldQuery
-
getBestFragment(FieldQuery, IndexReader, int, String, int): String
-
getBestFragments(FieldQuery, IndexReader, int, String, int, int): String[]
-
getBestFragment(FieldQuery, IndexReader, int, String, int, FragListBuilder, FragmentsBuilder, String[], String[], Encoder): String
-
getBestFragments(FieldQuery, IndexReader, int, String, int, int, FragListBuilder, FragmentsBuilder, String[], String[], Encoder): String[]
-
getBestFragments(FieldQuery, IndexReader, int, String, Set<String>, int, int, FragListBuilder, FragmentsBuilder, String[], String[], Encoder): String[]
-
getFieldFragList(FragListBuilder, FieldQuery, IndexReader, int, String, int): FieldFragList
-
getFieldFragList(FragListBuilder, FieldQuery, IndexReader, int, Set<String>, int): FieldFragList
-
isPhraseHighlight(): boolean
-
isFieldMatch(): boolean
-
getPhraseLimit(): int
-
setPhraseLimit(int): void
-
FieldFragList
-
FieldPhraseList
-
FieldQuery
-
fieldMatch: boolean
-
rootMaps: Map<String, QueryPhraseMap>
-
termSetMap: Map<String, Set<String>>
-
termOrPhraseNumber: int
-
MAX_MTQ_TERMS: int
-
FieldQuery(Query, IndexReader, boolean, boolean): void
-
FieldQuery(Query, boolean, boolean): void
-
flatten(Query, IndexReader, Collection<Query>, float): void
-
expand(Collection<Query>): Collection<Query>
-
checkOverlap(Collection<Query>, PhraseQuery, float, PhraseQuery, float): void
-
checkOverlap(Collection<Query>, Term[], Term[], int, float): void
-
getRootMap(Query): QueryPhraseMap
-
getKey(Query): String
-
saveTerms(Collection<Query>, IndexReader): void
-
getTermSet(Query): Set<String>
-
getTermSet(String): Set<String>
-
getFieldTermMap(String, String): QueryPhraseMap
-
searchPhrase(String, List<TermInfo>): QueryPhraseMap
-
getRootMap(String): QueryPhraseMap
-
nextTermOrPhraseNumber(): int
-
QueryPhraseMap
-
FieldTermStack
-
FragListBuilder
-
FragmentsBuilder
-
createFragment(IndexReader, int, String, FieldFragList): String
-
createFragments(IndexReader, int, String, FieldFragList, int): String[]
-
createFragment(IndexReader, int, String, FieldFragList, String[], String[], Encoder): String
-
createFragments(IndexReader, int, String, FieldFragList, int, String[], String[], Encoder): String[]
-
ScoreOrderFragmentsBuilder
-
SimpleBoundaryScanner
-
SimpleFieldFragList
-
SimpleFragListBuilder
-
SimpleFragmentsBuilder
-
SingleFragListBuilder
-
WeightedFieldFragList
-
WeightedFragListBuilder
Java files
Class files