/* * SPDX-License-Identifier: Apache-2.0 * * The OpenSearch Contributors require contributions made to * this file be licensed under the Apache-2.0 license or a * compatible open source license. */ /* * Licensed to Elasticsearch under one or more contributor * license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright * ownership. Elasticsearch licenses this file to you under * the Apache License, Version 2.0 (the "License"); you may * not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ /* * Modifications Copyright OpenSearch Contributors. See * GitHub history for details. */ package org.opensearch.search.aggregations.metrics; import org.HdrHistogram.DoubleHistogram; import org.opensearch.core.common.io.stream.StreamInput; import org.opensearch.core.common.io.stream.StreamOutput; import org.opensearch.core.xcontent.XContentBuilder; import org.opensearch.search.DocValueFormat; import org.opensearch.search.aggregations.InternalAggregation; import java.io.IOException; import java.nio.ByteBuffer; import java.util.Arrays; import java.util.List; import java.util.Map; import java.util.Objects; import java.util.zip.DataFormatException; /** * Base implementation of HDR percentiles * * @opensearch.internal */ abstract class AbstractInternalHDRPercentiles extends InternalNumericMetricsAggregation.MultiValue { protected final double[] keys; protected final DoubleHistogram state; protected final boolean keyed; AbstractInternalHDRPercentiles( String name, double[] keys, DoubleHistogram state, boolean keyed, DocValueFormat format, Map metadata ) { super(name, metadata); this.keys = keys; this.state = state; this.keyed = keyed; this.format = format; } /** * Read from a stream. */ protected AbstractInternalHDRPercentiles(StreamInput in) throws IOException { super(in); format = in.readNamedWriteable(DocValueFormat.class); keys = in.readDoubleArray(); long minBarForHighestToLowestValueRatio = in.readLong(); final int serializedLen = in.readVInt(); byte[] bytes = new byte[serializedLen]; in.readBytes(bytes, 0, serializedLen); ByteBuffer stateBuffer = ByteBuffer.wrap(bytes); try { state = DoubleHistogram.decodeFromCompressedByteBuffer(stateBuffer, minBarForHighestToLowestValueRatio); } catch (DataFormatException e) { throw new IOException("Failed to decode DoubleHistogram for aggregation [" + name + "]", e); } keyed = in.readBoolean(); } @Override protected void doWriteTo(StreamOutput out) throws IOException { out.writeNamedWriteable(format); out.writeDoubleArray(keys); out.writeLong(state.getHighestToLowestValueRatio()); ByteBuffer stateBuffer = ByteBuffer.allocate(state.getNeededByteBufferCapacity()); final int serializedLen = state.encodeIntoCompressedByteBuffer(stateBuffer); out.writeVInt(serializedLen); out.writeBytes(stateBuffer.array(), 0, serializedLen); out.writeBoolean(keyed); } @Override public double value(String name) { return value(Double.parseDouble(name)); } public DocValueFormat formatter() { return format; } public abstract double value(double key); public long getEstimatedMemoryFootprint() { return state.getEstimatedFootprintInBytes(); } /** * Return the internal {@link DoubleHistogram} sketch for this metric. */ public DoubleHistogram getState() { return state; } /** * Return the keys (percentiles) requested. */ public double[] getKeys() { return keys; } /** * Should the output be keyed. */ public boolean keyed() { return keyed; } @Override public AbstractInternalHDRPercentiles reduce(List aggregations, ReduceContext reduceContext) { DoubleHistogram merged = null; for (InternalAggregation aggregation : aggregations) { final AbstractInternalHDRPercentiles percentiles = (AbstractInternalHDRPercentiles) aggregation; if (merged == null) { merged = new DoubleHistogram(percentiles.state); merged.setAutoResize(true); } merged.add(percentiles.state); } return createReduced(getName(), keys, merged, keyed, getMetadata()); } protected abstract AbstractInternalHDRPercentiles createReduced( String name, double[] keys, DoubleHistogram merged, boolean keyed, Map metadata ); @Override public XContentBuilder doXContentBody(XContentBuilder builder, Params params) throws IOException { if (keyed) { builder.startObject(CommonFields.VALUES.getPreferredName()); for (int i = 0; i < keys.length; ++i) { String key = String.valueOf(keys[i]); double value = value(keys[i]); builder.field(key, state.getTotalCount() == 0 ? null : value); if (format != DocValueFormat.RAW && state.getTotalCount() > 0) { builder.field(key + "_as_string", format.format(value).toString()); } } builder.endObject(); } else { builder.startArray(CommonFields.VALUES.getPreferredName()); for (int i = 0; i < keys.length; i++) { double value = value(keys[i]); builder.startObject(); builder.field(CommonFields.KEY.getPreferredName(), keys[i]); builder.field(CommonFields.VALUE.getPreferredName(), state.getTotalCount() == 0 ? null : value); if (format != DocValueFormat.RAW && state.getTotalCount() > 0) { builder.field(CommonFields.VALUE_AS_STRING.getPreferredName(), format.format(value).toString()); } builder.endObject(); } builder.endArray(); } return builder; } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null || getClass() != obj.getClass()) return false; if (super.equals(obj) == false) return false; AbstractInternalHDRPercentiles that = (AbstractInternalHDRPercentiles) obj; return keyed == that.keyed && Arrays.equals(keys, that.keys) && Objects.equals(state, that.state); } @Override public int hashCode() { // we cannot use state.hashCode at the moment because of: // https://github.com/HdrHistogram/HdrHistogram/issues/81 // TODO: upgrade the HDRHistogram library return Objects.hash( super.hashCode(), keyed, Arrays.hashCode(keys), state.getIntegerToDoubleValueConversionRatio(), state.getTotalCount() ); } }