Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion specification/search/data-plane/Search/client.tsp
Original file line number Diff line number Diff line change
Expand Up @@ -2720,7 +2720,10 @@ namespace KnowledgeBaseRetrievalClient {
@@clientName(LuceneStandardTokenizer.`@odata.type`, "odataType");
@@clientName(LuceneStandardTokenizer.`@odata.type`, "odatatype", "javascript");
@@clientName(LuceneStandardTokenizerV2.`@odata.type`, "odataType");
@@clientName(LuceneStandardTokenizerV2.`@odata.type`, "odatatype", "javascript");
@@clientName(LuceneStandardTokenizerV2.`@odata.type`,
"odatatype",
"javascript"
);
@@clientName(UaxUrlEmailTokenizer.`@odata.type`, "odataType");
@@clientName(UaxUrlEmailTokenizer.`@odata.type`, "odatatype", "javascript");
@@clientName(AsciiFoldingTokenFilter.`@odata.type`, "odataType");
Expand Down
4 changes: 2 additions & 2 deletions specification/search/data-plane/Search/models-service.tsp
Original file line number Diff line number Diff line change
Expand Up @@ -5109,8 +5109,8 @@ model DictionaryDecompounderTokenFilter extends TokenFilter {
}

#suppress "@azure-tools/typespec-azure-core/casing-style" "Maintain compatibility with existing models."
@doc("Generates n-grams of the given size(s) starting from the front or the back of an input token. This token filter is implemented using Apache Lucene.")
#deprecated "EdgeNGramTokenFilter is deprecated. Use EdgeNGramTokenFilterV2 instead."
@doc("Generates n-grams of the given size(s) starting from the front or the back of an input token. This token filter is implemented using Apache Lucene.")
model EdgeNGramTokenFilter extends TokenFilter {
@doc("The minimum n-gram length. Default is 1. Must be less than the value of maxGram.")
minGram?: int32 = 1;
Expand Down Expand Up @@ -5210,8 +5210,8 @@ model LimitTokenFilter extends TokenFilter {
}

#suppress "@azure-tools/typespec-azure-core/casing-style" "Maintain compatibility with existing models."
@doc("Generates n-grams of the given size(s). This token filter is implemented using Apache Lucene.")
#deprecated "NGramTokenFilter is deprecated. Use NGramTokenFilterV2 instead."
@doc("Generates n-grams of the given size(s). This token filter is implemented using Apache Lucene.")
model NGramTokenFilter extends TokenFilter {
@doc("The minimum n-gram length. Default is 1. Must be less than the value of maxGram.")
minGram?: int32 = 1;
Expand Down
Loading