/* * SPDX-License-Identifier: Apache-2.0 * * The OpenSearch Contributors require contributions made to * this file be licensed under the Apache-2.0 license or a * compatible open source license. */ /* * Licensed to Elasticsearch under one or more contributor * license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright * ownership. Elasticsearch licenses this file to you under * the Apache License, Version 2.0 (the "License"); you may * not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ /* * Modifications Copyright OpenSearch Contributors. See * GitHub history for details. */ package org.opensearch.search.aggregations; import org.apache.lucene.search.LeafCollector; import org.apache.lucene.search.Scorable; import org.opensearch.search.aggregations.bucket.terms.LongKeyedBucketOrds; import java.io.IOException; import java.util.stream.Stream; import java.util.stream.StreamSupport; /** * Per-leaf bucket collector. * * @opensearch.internal */ public abstract class LeafBucketCollector implements LeafCollector { public static final LeafBucketCollector NO_OP_COLLECTOR = new LeafBucketCollector() { @Override public void setScorer(Scorable arg0) throws IOException { // no-op } @Override public void collect(int doc, long bucket) { // no-op } }; public static LeafBucketCollector wrap(Iterable collectors) { final Stream actualCollectors = StreamSupport.stream(collectors.spliterator(), false) .filter(c -> c != NO_OP_COLLECTOR); final LeafBucketCollector[] colls = actualCollectors.toArray(size -> new LeafBucketCollector[size]); switch (colls.length) { case 0: return NO_OP_COLLECTOR; case 1: return colls[0]; default: return new LeafBucketCollector() { @Override public void setScorer(Scorable s) throws IOException { for (LeafBucketCollector c : colls) { c.setScorer(s); } } @Override public void collect(int doc, long bucket) throws IOException { for (LeafBucketCollector c : colls) { c.collect(doc, bucket); } } }; } } /** * Collect the given {@code doc} in the bucket owned by * {@code owningBucketOrd}. *

* The implementation of this method metric aggregations is generally * something along the lines of *

{@code
     * array[owningBucketOrd] += loadValueFromDoc(doc)
     * }
*

Bucket aggregations have more trouble because their job is to * make new ordinals. So their implementation generally * looks kind of like *

{@code
     * long myBucketOrd = mapOwningBucketAndValueToMyOrd(owningBucketOrd, loadValueFromDoc(doc));
     * collectBucket(doc, myBucketOrd);
     * }
*

* Some bucket aggregations "know" how many ordinals each owning ordinal * needs so they can map "densely". The {@code range} aggregation, for * example, can perform this mapping with something like: *

{@code
     * return rangeCount * owningBucketOrd + matchingRange(value);
     * }
* Other aggregations don't know how many buckets will fall into any * particular owning bucket. The {@code terms} aggregation, for example, * uses {@link LongKeyedBucketOrds} which amounts to a hash lookup. */ public abstract void collect(int doc, long owningBucketOrd) throws IOException; @Override public final void collect(int doc) throws IOException { collect(doc, 0); } @Override public void setScorer(Scorable scorer) throws IOException { // no-op by default } }