/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.lucene.codecs;


import java.io.Closeable;
import java.io.IOException;

import org.apache.lucene.index.FieldInfo;
import org.apache.lucene.index.MergeState;
import org.apache.lucene.index.PointValues;

Abstract API to write points
@lucene.experimental
/** Abstract API to write points * * @lucene.experimental */
public abstract class PointsWriter implements Closeable {
Sole constructor. (For invocation by subclass constructors, typically implicit.)
/** Sole constructor. (For invocation by subclass * constructors, typically implicit.) */
protected PointsWriter() { }
Write all values contained in the provided reader
/** Write all values contained in the provided reader */
public abstract void writeField(FieldInfo fieldInfo, PointsReader values) throws IOException;
Default naive merge implementation for one field: it just re-indexes all the values from the incoming segment. The default codec overrides this for 1D fields and uses a faster but more complex implementation.
/** Default naive merge implementation for one field: it just re-indexes all the values * from the incoming segment. The default codec overrides this for 1D fields and uses * a faster but more complex implementation. */
protected void mergeOneField(MergeState mergeState, FieldInfo fieldInfo) throws IOException { long maxPointCount = 0; int docCount = 0; for (int i=0;i<mergeState.pointsReaders.length;i++) { PointsReader pointsReader = mergeState.pointsReaders[i]; if (pointsReader != null) { FieldInfo readerFieldInfo = mergeState.fieldInfos[i].fieldInfo(fieldInfo.name); if (readerFieldInfo != null && readerFieldInfo.getPointDataDimensionCount() > 0) { PointValues values = pointsReader.getValues(fieldInfo.name); if (values != null) { maxPointCount += values.size(); docCount += values.getDocCount(); } } } } final long finalMaxPointCount = maxPointCount; final int finalDocCount = docCount; writeField(fieldInfo, new PointsReader() { @Override public long ramBytesUsed() { return 0; } @Override public void close() throws IOException {} @Override public PointValues getValues(String fieldName) { if (fieldName.equals(fieldInfo.name) == false) { throw new IllegalArgumentException("field name must match the field being merged"); } return new PointValues() { @Override public void intersect(IntersectVisitor mergedVisitor) throws IOException { for (int i=0;i<mergeState.pointsReaders.length;i++) { PointsReader pointsReader = mergeState.pointsReaders[i]; if (pointsReader == null) { // This segment has no points continue; } FieldInfo readerFieldInfo = mergeState.fieldInfos[i].fieldInfo(fieldName); if (readerFieldInfo == null) { // This segment never saw this field continue; } if (readerFieldInfo.getPointDataDimensionCount() == 0) { // This segment saw this field, but the field did not index points in it: continue; } PointValues values = pointsReader.getValues(fieldName); if (values == null) { continue; } MergeState.DocMap docMap = mergeState.docMaps[i]; values.intersect(new IntersectVisitor() { @Override public void visit(int docID) { // Should never be called because our compare method never returns Relation.CELL_INSIDE_QUERY throw new IllegalStateException(); } @Override public void visit(int docID, byte[] packedValue) throws IOException { int newDocID = docMap.get(docID); if (newDocID != -1) { // Not deleted: mergedVisitor.visit(newDocID, packedValue); } } @Override public Relation compare(byte[] minPackedValue, byte[] maxPackedValue) { // Forces this segment's PointsReader to always visit all docs + values: return Relation.CELL_CROSSES_QUERY; } }); } } @Override public long estimatePointCount(IntersectVisitor visitor) { throw new UnsupportedOperationException(); } @Override public byte[] getMinPackedValue() { throw new UnsupportedOperationException(); } @Override public byte[] getMaxPackedValue() { throw new UnsupportedOperationException(); } @Override public int getNumDataDimensions() { throw new UnsupportedOperationException(); } @Override public int getNumIndexDimensions() { throw new UnsupportedOperationException(); } @Override public int getBytesPerDimension() { throw new UnsupportedOperationException(); } @Override public long size() { return finalMaxPointCount; } @Override public int getDocCount() { return finalDocCount; } }; } @Override public void checkIntegrity() throws IOException { throw new UnsupportedOperationException(); } }); }
Default merge implementation to merge incoming points readers by visiting all their points and adding to this writer
/** Default merge implementation to merge incoming points readers by visiting all their points and * adding to this writer */
public void merge(MergeState mergeState) throws IOException { // check each incoming reader for (PointsReader reader : mergeState.pointsReaders) { if (reader != null) { reader.checkIntegrity(); } } // merge field at a time for (FieldInfo fieldInfo : mergeState.mergeFieldInfos) { if (fieldInfo.getPointDataDimensionCount() != 0) { mergeOneField(mergeState, fieldInfo); } } finish(); }
Called once at the end before close
/** Called once at the end before close */
public abstract void finish() throws IOException; }