/* SPDX-License-Identifier: Apache-2.0
*
* The OpenSearch Contributors require contributions made to
* this file be licensed under the Apache-2.0 license or a
* compatible open source license.
*/
/*
* Modifications Copyright OpenSearch Contributors. See
* GitHub history for details.
*
* Licensed to Elasticsearch B.V. under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch B.V. licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
using System.Collections.Generic;
using System.Runtime.Serialization;
using OpenSearch.Net.Utf8Json;
namespace OpenSearch.Client
{
///
/// A tokenizer of type edgeNGram.
///
public interface IEdgeNGramTokenizer : ITokenizer
{
///
/// Maximum size in codepoints of a single n-gram
///
[DataMember(Name ="max_gram")]
[JsonFormatter(typeof(NullableStringIntFormatter))]
int? MaxGram { get; set; }
///
/// Minimum size in codepoints of a single n-gram
///
[DataMember(Name ="min_gram")]
[JsonFormatter(typeof(NullableStringIntFormatter))]
int? MinGram { get; set; }
///
/// Characters classes to keep in the tokens, OpenSearch
/// will split on characters that don’t belong to any of these classes.
///
[DataMember(Name ="token_chars")]
IEnumerable TokenChars { get; set; }
///
/// Custom characters that should be treated as part of a token. For example,
/// setting this to +-_ will make the tokenizer treat the plus, minus and
/// underscore sign as part of a token.
///
/// Requires setting as part of
///
[DataMember(Name = "custom_token_chars")]
string CustomTokenChars { get; set; }
}
///
public class EdgeNGramTokenizer : TokenizerBase, IEdgeNGramTokenizer
{
public EdgeNGramTokenizer() => Type = "edge_ngram";
///
public int? MaxGram { get; set; }
///
public int? MinGram { get; set; }
///
public IEnumerable TokenChars { get; set; }
///
public string CustomTokenChars { get; set; }
}
///
public class EdgeNGramTokenizerDescriptor
: TokenizerDescriptorBase, IEdgeNGramTokenizer
{
protected override string Type => "edge_ngram";
int? IEdgeNGramTokenizer.MaxGram { get; set; }
int? IEdgeNGramTokenizer.MinGram { get; set; }
IEnumerable IEdgeNGramTokenizer.TokenChars { get; set; }
string IEdgeNGramTokenizer.CustomTokenChars { get; set; }
///
public EdgeNGramTokenizerDescriptor MinGram(int? minGram) => Assign(minGram, (a, v) => a.MinGram = v);
///
public EdgeNGramTokenizerDescriptor MaxGram(int? maxGram) => Assign(maxGram, (a, v) => a.MaxGram = v);
///
public EdgeNGramTokenizerDescriptor TokenChars(IEnumerable tokenChars) =>
Assign(tokenChars, (a, v) => a.TokenChars = v);
///
public EdgeNGramTokenizerDescriptor TokenChars(params TokenChar[] tokenChars) =>
Assign(tokenChars, (a, v) => a.TokenChars = v);
///
public EdgeNGramTokenizerDescriptor CustomTokenChars(string customTokenChars) =>
Assign(customTokenChars, (a, v) => a.CustomTokenChars = v);
}
}