From 04bcd64708669ab2d50a5b40e727b97be6580def Mon Sep 17 00:00:00 2001 From: Sima Zhu Date: Fri, 29 May 2020 11:09:55 -0700 Subject: [PATCH 1/4] Collapse multiple versions into one. --- .../converters/BM25SimilarityConverter.java | 14 +-- .../ClassicSimilarityConverter.java | 14 +-- .../EdgeNGramTokenFilterConverter.java | 34 +++++- .../EdgeNGramTokenFilterV2Converter.java | 70 ------------ .../converters/KeywordTokenizerConverter.java | 40 +++++-- .../KeywordTokenizerV2Converter.java | 52 --------- .../converters/LexicalTokenizerConverter.java | 16 ++- .../LuceneStandardTokenizerConverter.java | 26 ++++- .../LuceneStandardTokenizerV2Converter.java | 52 --------- .../converters/NGramTokenFilterConverter.java | 28 ++++- .../NGramTokenFilterV2Converter.java | 58 ---------- ...a => PathHierarchyTokenizerConverter.java} | 32 +++--- .../converters/SearchIndexConverter.java | 20 ++-- .../converters/SimilarityConverter.java | 16 +-- .../converters/SuggesterConverter.java | 20 ++-- .../converters/TokenFilterConverter.java | 1 - .../models/EdgeNGramTokenFilter.java | 2 +- ...rity.java => BM25SimilarityAlgorithm.java} | 6 +- ...y.java => ClassicSimilarityAlgorithm.java} | 2 +- .../models/EdgeNGramTokenFilterV2.java | 106 ------------------ .../indexes/models/KeywordTokenizer.java | 34 +++--- .../indexes/models/KeywordTokenizerV2.java | 50 --------- .../indexes/models/LexicalTokenizer.java | 4 +- .../models/LuceneStandardTokenizerV2.java | 50 --------- .../indexes/models/NGramTokenFilterV2.java | 75 ------------- ...zerV2.java => PathHierarchyTokenizer.java} | 12 +- .../documents/indexes/models/SearchIndex.java | 24 ++-- .../{Suggester.java => SearchSuggester.java} | 6 +- ...milarity.java => SimilarityAlgorithm.java} | 8 +- .../IndexAndServiceStatisticsExample.java | 4 +- .../indexes/LifecycleSetupExample.java | 4 +- .../search/documents/SearchTestBase.java | 4 +- .../indexes/CustomAnalyzerSyncTests.java | 19 ++-- .../indexes/IndexManagementSyncTests.java | 6 +- 34 files changed, 239 insertions(+), 670 deletions(-) delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterV2Converter.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerV2Converter.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerV2Converter.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterV2Converter.java rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/{PathHierarchyTokenizerV2Converter.java => PathHierarchyTokenizerConverter.java} (68%) rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/{BM25Similarity.java => BM25SimilarityAlgorithm.java} (94%) rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/{ClassicSimilarity.java => ClassicSimilarityAlgorithm.java} (90%) delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/EdgeNGramTokenFilterV2.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizerV2.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LuceneStandardTokenizerV2.java delete mode 100644 sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/NGramTokenFilterV2.java rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/{PathHierarchyTokenizerV2.java => PathHierarchyTokenizer.java} (90%) rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/{Suggester.java => SearchSuggester.java} (92%) rename sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/{Similarity.java => SimilarityAlgorithm.java} (83%) diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/BM25SimilarityConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/BM25SimilarityConverter.java index 81d62afc2ea6..5a9a92b81298 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/BM25SimilarityConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/BM25SimilarityConverter.java @@ -3,21 +3,21 @@ package com.azure.search.documents.implementation.converters; -import com.azure.search.documents.indexes.models.BM25Similarity; +import com.azure.search.documents.indexes.models.BM25SimilarityAlgorithm; /** * A converter between {@link com.azure.search.documents.indexes.implementation.models.BM25Similarity} and - * {@link BM25Similarity}. + * {@link BM25SimilarityAlgorithm}. */ public final class BM25SimilarityConverter { /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.BM25Similarity} to {@link BM25Similarity}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.BM25Similarity} to {@link BM25SimilarityAlgorithm}. */ - public static BM25Similarity map(com.azure.search.documents.indexes.implementation.models.BM25Similarity obj) { + public static BM25SimilarityAlgorithm map(com.azure.search.documents.indexes.implementation.models.BM25Similarity obj) { if (obj == null) { return null; } - BM25Similarity bM25Similarity = new BM25Similarity(); + BM25SimilarityAlgorithm bM25Similarity = new BM25SimilarityAlgorithm(); Double b = obj.getB(); bM25Similarity.setB(b); @@ -28,9 +28,9 @@ public static BM25Similarity map(com.azure.search.documents.indexes.implementati } /** - * Maps from {@link BM25Similarity} to {@link com.azure.search.documents.indexes.implementation.models.BM25Similarity}. + * Maps from {@link BM25SimilarityAlgorithm} to {@link com.azure.search.documents.indexes.implementation.models.BM25Similarity}. */ - public static com.azure.search.documents.indexes.implementation.models.BM25Similarity map(BM25Similarity obj) { + public static com.azure.search.documents.indexes.implementation.models.BM25Similarity map(BM25SimilarityAlgorithm obj) { if (obj == null) { return null; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/ClassicSimilarityConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/ClassicSimilarityConverter.java index 1884baa77242..0dcbd8f3f647 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/ClassicSimilarityConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/ClassicSimilarityConverter.java @@ -3,30 +3,30 @@ package com.azure.search.documents.implementation.converters; -import com.azure.search.documents.indexes.models.ClassicSimilarity; +import com.azure.search.documents.indexes.models.ClassicSimilarityAlgorithm; /** * A converter between {@link com.azure.search.documents.indexes.implementation.models.ClassicSimilarity} and - * {@link ClassicSimilarity}. + * {@link ClassicSimilarityAlgorithm}. */ public final class ClassicSimilarityConverter { /** * Maps from {@link com.azure.search.documents.indexes.implementation.models.ClassicSimilarity} to - * {@link ClassicSimilarity}. + * {@link ClassicSimilarityAlgorithm}. */ - public static ClassicSimilarity map(com.azure.search.documents.indexes.implementation.models.ClassicSimilarity obj) { + public static ClassicSimilarityAlgorithm map(com.azure.search.documents.indexes.implementation.models.ClassicSimilarity obj) { if (obj == null) { return null; } - ClassicSimilarity classicSimilarity = new ClassicSimilarity(); + ClassicSimilarityAlgorithm classicSimilarity = new ClassicSimilarityAlgorithm(); return classicSimilarity; } /** - * Maps from {@link ClassicSimilarity} to + * Maps from {@link ClassicSimilarityAlgorithm} to * {@link com.azure.search.documents.indexes.implementation.models.ClassicSimilarity}. */ - public static com.azure.search.documents.indexes.implementation.models.ClassicSimilarity map(ClassicSimilarity obj) { + public static com.azure.search.documents.indexes.implementation.models.ClassicSimilarity map(ClassicSimilarityAlgorithm obj) { if (obj == null) { return null; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterConverter.java index 21f7af6e9526..1635b8c717b6 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterConverter.java @@ -37,16 +37,42 @@ public static EdgeNGramTokenFilter map(com.azure.search.documents.indexes.implem return edgeNGramTokenFilter; } + /** + * Maps from {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2} to + * {@link EdgeNGramTokenFilter}. + */ + public static EdgeNGramTokenFilter map(com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 obj) { + if (obj == null) { + return null; + } + EdgeNGramTokenFilter edgeNGramTokenFilter = new EdgeNGramTokenFilter(); + + String name = obj.getName(); + edgeNGramTokenFilter.setName(name); + + Integer maxGram = obj.getMaxGram(); + edgeNGramTokenFilter.setMaxGram(maxGram); + + if (obj.getSide() != null) { + EdgeNGramTokenFilterSide side = EdgeNGramTokenFilterSideConverter.map(obj.getSide()); + edgeNGramTokenFilter.setSide(side); + } + + Integer minGram = obj.getMinGram(); + edgeNGramTokenFilter.setMinGram(minGram); + return edgeNGramTokenFilter; + } + /** * Maps from {@link EdgeNGramTokenFilter} to - * {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter}. + * {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2}. */ - public static com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter map(EdgeNGramTokenFilter obj) { + public static com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 map(EdgeNGramTokenFilter obj) { if (obj == null) { return null; } - com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter edgeNGramTokenFilter = - new com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter(); + com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 edgeNGramTokenFilter = + new com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2(); String name = obj.getName(); edgeNGramTokenFilter.setName(name); diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterV2Converter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterV2Converter.java deleted file mode 100644 index bdb2a3c6b54d..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/EdgeNGramTokenFilterV2Converter.java +++ /dev/null @@ -1,70 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.implementation.converters; - -import com.azure.search.documents.indexes.models.EdgeNGramTokenFilterSide; -import com.azure.search.documents.indexes.models.EdgeNGramTokenFilterV2; - -/** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2} and - * {@link EdgeNGramTokenFilterV2}. - */ -public final class EdgeNGramTokenFilterV2Converter { - /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2} to - * {@link EdgeNGramTokenFilterV2}. - */ - public static EdgeNGramTokenFilterV2 map(com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 obj) { - if (obj == null) { - return null; - } - EdgeNGramTokenFilterV2 edgeNGramTokenFilterV2 = new EdgeNGramTokenFilterV2(); - - String name = obj.getName(); - edgeNGramTokenFilterV2.setName(name); - - Integer maxGram = obj.getMaxGram(); - edgeNGramTokenFilterV2.setMaxGram(maxGram); - - if (obj.getSide() != null) { - EdgeNGramTokenFilterSide side = EdgeNGramTokenFilterSideConverter.map(obj.getSide()); - edgeNGramTokenFilterV2.setSide(side); - } - - Integer minGram = obj.getMinGram(); - edgeNGramTokenFilterV2.setMinGram(minGram); - return edgeNGramTokenFilterV2; - } - - /** - * Maps from {@link EdgeNGramTokenFilterV2} to - * {@link com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2}. - */ - public static com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 map(EdgeNGramTokenFilterV2 obj) { - if (obj == null) { - return null; - } - com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2 edgeNGramTokenFilterV2 = - new com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2(); - - String name = obj.getName(); - edgeNGramTokenFilterV2.setName(name); - - Integer maxGram = obj.getMaxGram(); - edgeNGramTokenFilterV2.setMaxGram(maxGram); - - if (obj.getSide() != null) { - com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterSide side = - EdgeNGramTokenFilterSideConverter.map(obj.getSide()); - edgeNGramTokenFilterV2.setSide(side); - } - - Integer minGram = obj.getMinGram(); - edgeNGramTokenFilterV2.setMinGram(minGram); - return edgeNGramTokenFilterV2; - } - - private EdgeNGramTokenFilterV2Converter() { - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerConverter.java index c129572c4bfb..b13c060a6326 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerConverter.java @@ -6,14 +6,15 @@ import com.azure.search.documents.indexes.models.KeywordTokenizer; /** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizer} and + * A converter between {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2} and * {@link KeywordTokenizer}. */ public final class KeywordTokenizerConverter { /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizer} to {@link KeywordTokenizer}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2} to + * {@link KeywordTokenizer}. */ - public static KeywordTokenizer map(com.azure.search.documents.indexes.implementation.models.KeywordTokenizer obj) { + public static KeywordTokenizer map(com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 obj) { if (obj == null) { return null; } @@ -22,29 +23,48 @@ public static KeywordTokenizer map(com.azure.search.documents.indexes.implementa String name = obj.getName(); keywordTokenizer.setName(name); - Integer bufferSize = obj.getBufferSize(); - keywordTokenizer.setBufferSize(bufferSize); + Integer maxTokenLength = obj.getMaxTokenLength(); + keywordTokenizer.setMaxTokenLength(maxTokenLength); return keywordTokenizer; } /** - * Maps from {@link KeywordTokenizer} to {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizer}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizer} to + * {@link KeywordTokenizer}. */ - public static com.azure.search.documents.indexes.implementation.models.KeywordTokenizer map(KeywordTokenizer obj) { + public static KeywordTokenizer map(com.azure.search.documents.indexes.implementation.models.KeywordTokenizer obj) { if (obj == null) { return null; } - com.azure.search.documents.indexes.implementation.models.KeywordTokenizer keywordTokenizer = - new com.azure.search.documents.indexes.implementation.models.KeywordTokenizer(); + KeywordTokenizer keywordTokenizer = new KeywordTokenizer(); String name = obj.getName(); keywordTokenizer.setName(name); Integer bufferSize = obj.getBufferSize(); - keywordTokenizer.setBufferSize(bufferSize); + keywordTokenizer.setMaxTokenLength(bufferSize); return keywordTokenizer; } + /** + * Maps from {@link KeywordTokenizer} to + * {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2}. + */ + public static com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 map(KeywordTokenizer obj) { + if (obj == null) { + return null; + } + com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 keywordTokenizerV2 = + new com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2(); + + String name = obj.getName(); + keywordTokenizerV2.setName(name); + + Integer maxTokenLength = obj.getMaxTokenLength(); + keywordTokenizerV2.setMaxTokenLength(maxTokenLength); + return keywordTokenizerV2; + } + private KeywordTokenizerConverter() { } } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerV2Converter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerV2Converter.java deleted file mode 100644 index a39b50395a31..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/KeywordTokenizerV2Converter.java +++ /dev/null @@ -1,52 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.implementation.converters; - -import com.azure.search.documents.indexes.models.KeywordTokenizerV2; - -/** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2} and - * {@link KeywordTokenizerV2}. - */ -public final class KeywordTokenizerV2Converter { - /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2} to - * {@link KeywordTokenizerV2}. - */ - public static KeywordTokenizerV2 map(com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 obj) { - if (obj == null) { - return null; - } - KeywordTokenizerV2 keywordTokenizerV2 = new KeywordTokenizerV2(); - - String name = obj.getName(); - keywordTokenizerV2.setName(name); - - Integer maxTokenLength = obj.getMaxTokenLength(); - keywordTokenizerV2.setMaxTokenLength(maxTokenLength); - return keywordTokenizerV2; - } - - /** - * Maps from {@link KeywordTokenizerV2} to - * {@link com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2}. - */ - public static com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 map(KeywordTokenizerV2 obj) { - if (obj == null) { - return null; - } - com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2 keywordTokenizerV2 = - new com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2(); - - String name = obj.getName(); - keywordTokenizerV2.setName(name); - - Integer maxTokenLength = obj.getMaxTokenLength(); - keywordTokenizerV2.setMaxTokenLength(maxTokenLength); - return keywordTokenizerV2; - } - - private KeywordTokenizerV2Converter() { - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java index dc04bdab01a0..c5a46773e22e 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java @@ -7,13 +7,11 @@ import com.azure.search.documents.indexes.models.ClassicTokenizer; import com.azure.search.documents.indexes.models.EdgeNGramTokenizer; import com.azure.search.documents.indexes.models.KeywordTokenizer; -import com.azure.search.documents.indexes.models.KeywordTokenizerV2; import com.azure.search.documents.indexes.models.LuceneStandardTokenizer; -import com.azure.search.documents.indexes.models.LuceneStandardTokenizerV2; import com.azure.search.documents.indexes.models.MicrosoftLanguageStemmingTokenizer; import com.azure.search.documents.indexes.models.MicrosoftLanguageTokenizer; import com.azure.search.documents.indexes.models.NGramTokenizer; -import com.azure.search.documents.indexes.models.PathHierarchyTokenizerV2; +import com.azure.search.documents.indexes.models.PathHierarchyTokenizer; import com.azure.search.documents.indexes.models.PatternTokenizer; import com.azure.search.documents.indexes.models.UaxUrlEmailTokenizer; import com.azure.search.documents.indexes.models.LexicalTokenizer; @@ -43,7 +41,7 @@ public static LexicalTokenizer map(com.azure.search.documents.indexes.implementa return LuceneStandardTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizer) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2) { - return PathHierarchyTokenizerV2Converter.map((com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2) obj); + return PathHierarchyTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.ClassicTokenizer) { return ClassicTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.ClassicTokenizer) obj); @@ -58,7 +56,7 @@ public static LexicalTokenizer map(com.azure.search.documents.indexes.implementa return UaxUrlEmailTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.UaxUrlEmailTokenizer) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2) { - return KeywordTokenizerV2Converter.map((com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2) obj); + return KeywordTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.KeywordTokenizerV2) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.MicrosoftLanguageTokenizer) { return MicrosoftLanguageTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.MicrosoftLanguageTokenizer) obj); @@ -87,8 +85,8 @@ public static com.azure.search.documents.indexes.implementation.models.LexicalTo if (obj instanceof MicrosoftLanguageStemmingTokenizer) { return MicrosoftLanguageStemmingTokenizerConverter.map((MicrosoftLanguageStemmingTokenizer) obj); } - if (obj instanceof KeywordTokenizerV2) { - return KeywordTokenizerV2Converter.map((KeywordTokenizerV2) obj); + if (obj instanceof KeywordTokenizer) { + return KeywordTokenizerConverter.map((KeywordTokenizer) obj); } if (obj instanceof KeywordTokenizer) { return KeywordTokenizerConverter.map((KeywordTokenizer) obj); @@ -105,8 +103,8 @@ public static com.azure.search.documents.indexes.implementation.models.LexicalTo if (obj instanceof UaxUrlEmailTokenizer) { return UaxUrlEmailTokenizerConverter.map((UaxUrlEmailTokenizer) obj); } - if (obj instanceof PathHierarchyTokenizerV2) { - return PathHierarchyTokenizerV2Converter.map((PathHierarchyTokenizerV2) obj); + if (obj instanceof PathHierarchyTokenizer) { + return PathHierarchyTokenizerConverter.map((PathHierarchyTokenizer) obj); } if (obj instanceof ClassicTokenizer) { return ClassicTokenizerConverter.map((ClassicTokenizer) obj); diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerConverter.java index 99be37f5f562..524741635e16 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerConverter.java @@ -28,16 +28,34 @@ public static LuceneStandardTokenizer map(com.azure.search.documents.indexes.imp return luceneStandardTokenizer; } + /** + * Maps from {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2} to + * {@link LuceneStandardTokenizer}. + */ + public static LuceneStandardTokenizer map(com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 obj) { + if (obj == null) { + return null; + } + LuceneStandardTokenizer luceneStandardTokenizer = new LuceneStandardTokenizer(); + + String name = obj.getName(); + luceneStandardTokenizer.setName(name); + + Integer maxTokenLength = obj.getMaxTokenLength(); + luceneStandardTokenizer.setMaxTokenLength(maxTokenLength); + return luceneStandardTokenizer; + } + /** * Maps from {@link LuceneStandardTokenizer} to - * {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizer}. + * {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2}. */ - public static com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizer map(LuceneStandardTokenizer obj) { + public static com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 map(LuceneStandardTokenizer obj) { if (obj == null) { return null; } - com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizer luceneStandardTokenizer = - new com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizer(); + com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 luceneStandardTokenizer = + new com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2(); String name = obj.getName(); luceneStandardTokenizer.setName(name); diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerV2Converter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerV2Converter.java deleted file mode 100644 index 5fc6a4229474..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LuceneStandardTokenizerV2Converter.java +++ /dev/null @@ -1,52 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.implementation.converters; - -import com.azure.search.documents.indexes.models.LuceneStandardTokenizerV2; - -/** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2} and - * {@link LuceneStandardTokenizerV2}. - */ -public final class LuceneStandardTokenizerV2Converter { - /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2} to - * {@link LuceneStandardTokenizerV2}. - */ - public static LuceneStandardTokenizerV2 map(com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 obj) { - if (obj == null) { - return null; - } - LuceneStandardTokenizerV2 luceneStandardTokenizerV2 = new LuceneStandardTokenizerV2(); - - String name = obj.getName(); - luceneStandardTokenizerV2.setName(name); - - Integer maxTokenLength = obj.getMaxTokenLength(); - luceneStandardTokenizerV2.setMaxTokenLength(maxTokenLength); - return luceneStandardTokenizerV2; - } - - /** - * Maps from {@link LuceneStandardTokenizerV2} to - * {@link com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2}. - */ - public static com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 map(LuceneStandardTokenizerV2 obj) { - if (obj == null) { - return null; - } - com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2 luceneStandardTokenizerV2 = - new com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2(); - - String name = obj.getName(); - luceneStandardTokenizerV2.setName(name); - - Integer maxTokenLength = obj.getMaxTokenLength(); - luceneStandardTokenizerV2.setMaxTokenLength(maxTokenLength); - return luceneStandardTokenizerV2; - } - - private LuceneStandardTokenizerV2Converter() { - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterConverter.java index 0ac0f7d2246c..aaeb3d1dae4a 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterConverter.java @@ -31,14 +31,34 @@ public static NGramTokenFilter map(com.azure.search.documents.indexes.implementa } /** - * Maps from {@link NGramTokenFilter} to {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilter}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2} to {@link NGramTokenFilter}. */ - public static com.azure.search.documents.indexes.implementation.models.NGramTokenFilter map(NGramTokenFilter obj) { + public static NGramTokenFilter map(com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 obj) { if (obj == null) { return null; } - com.azure.search.documents.indexes.implementation.models.NGramTokenFilter nGramTokenFilter = - new com.azure.search.documents.indexes.implementation.models.NGramTokenFilter(); + NGramTokenFilter nGramTokenFilter = new NGramTokenFilter(); + + String name = obj.getName(); + nGramTokenFilter.setName(name); + + Integer maxGram = obj.getMaxGram(); + nGramTokenFilter.setMaxGram(maxGram); + + Integer minGram = obj.getMinGram(); + nGramTokenFilter.setMinGram(minGram); + return nGramTokenFilter; + } + + /** + * Maps from {@link NGramTokenFilter} to {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2}. + */ + public static com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 map(NGramTokenFilter obj) { + if (obj == null) { + return null; + } + com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 nGramTokenFilter = + new com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2(); String name = obj.getName(); nGramTokenFilter.setName(name); diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterV2Converter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterV2Converter.java deleted file mode 100644 index ef2e66bab3ef..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/NGramTokenFilterV2Converter.java +++ /dev/null @@ -1,58 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.implementation.converters; - -import com.azure.search.documents.indexes.models.NGramTokenFilterV2; - -/** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2} and - * {@link NGramTokenFilterV2}. - */ -public final class NGramTokenFilterV2Converter { - /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2} to - * {@link NGramTokenFilterV2}. - */ - public static NGramTokenFilterV2 map(com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 obj) { - if (obj == null) { - return null; - } - NGramTokenFilterV2 nGramTokenFilterV2 = new NGramTokenFilterV2(); - - String name = obj.getName(); - nGramTokenFilterV2.setName(name); - - Integer maxGram = obj.getMaxGram(); - nGramTokenFilterV2.setMaxGram(maxGram); - - Integer minGram = obj.getMinGram(); - nGramTokenFilterV2.setMinGram(minGram); - return nGramTokenFilterV2; - } - - /** - * Maps from {@link NGramTokenFilterV2} to - * {@link com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2}. - */ - public static com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 map(NGramTokenFilterV2 obj) { - if (obj == null) { - return null; - } - com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2 nGramTokenFilterV2 = - new com.azure.search.documents.indexes.implementation.models.NGramTokenFilterV2(); - - String name = obj.getName(); - nGramTokenFilterV2.setName(name); - - Integer maxGram = obj.getMaxGram(); - nGramTokenFilterV2.setMaxGram(maxGram); - - Integer minGram = obj.getMinGram(); - nGramTokenFilterV2.setMinGram(minGram); - return nGramTokenFilterV2; - } - - private NGramTokenFilterV2Converter() { - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerV2Converter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerConverter.java similarity index 68% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerV2Converter.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerConverter.java index 16b511a5db09..d68ba77066f3 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerV2Converter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/PathHierarchyTokenizerConverter.java @@ -3,48 +3,48 @@ package com.azure.search.documents.implementation.converters; -import com.azure.search.documents.indexes.models.PathHierarchyTokenizerV2; +import com.azure.search.documents.indexes.models.PathHierarchyTokenizer; /** * A converter between {@link com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2} and - * {@link PathHierarchyTokenizerV2}. + * {@link PathHierarchyTokenizer}. */ -public final class PathHierarchyTokenizerV2Converter { +public final class PathHierarchyTokenizerConverter { /** * Maps from {@link com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2} to - * {@link PathHierarchyTokenizerV2}. + * {@link PathHierarchyTokenizer}. */ - public static PathHierarchyTokenizerV2 map(com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2 obj) { + public static PathHierarchyTokenizer map(com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2 obj) { if (obj == null) { return null; } - PathHierarchyTokenizerV2 pathHierarchyTokenizerV2 = new PathHierarchyTokenizerV2(); + PathHierarchyTokenizer pathHierarchyTokenizer = new PathHierarchyTokenizer(); String name = obj.getName(); - pathHierarchyTokenizerV2.setName(name); + pathHierarchyTokenizer.setName(name); Integer maxTokenLength = obj.getMaxTokenLength(); - pathHierarchyTokenizerV2.setMaxTokenLength(maxTokenLength); + pathHierarchyTokenizer.setMaxTokenLength(maxTokenLength); String delimiter = obj.getDelimiter(); - pathHierarchyTokenizerV2.setDelimiter(delimiter); + pathHierarchyTokenizer.setDelimiter(delimiter); Boolean reverseTokenOrder = obj.isReverseTokenOrder(); - pathHierarchyTokenizerV2.setReverseTokenOrder(reverseTokenOrder); + pathHierarchyTokenizer.setReverseTokenOrder(reverseTokenOrder); Integer numberOfTokensToSkip = obj.getNumberOfTokensToSkip(); - pathHierarchyTokenizerV2.setNumberOfTokensToSkip(numberOfTokensToSkip); + pathHierarchyTokenizer.setNumberOfTokensToSkip(numberOfTokensToSkip); String replacement = obj.getReplacement(); - pathHierarchyTokenizerV2.setReplacement(replacement); - return pathHierarchyTokenizerV2; + pathHierarchyTokenizer.setReplacement(replacement); + return pathHierarchyTokenizer; } /** - * Maps from {@link PathHierarchyTokenizerV2} to + * Maps from {@link PathHierarchyTokenizer} to * {@link com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2}. */ - public static com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2 map(PathHierarchyTokenizerV2 obj) { + public static com.azure.search.documents.indexes.implementation.models.PathHierarchyTokenizerV2 map(PathHierarchyTokenizer obj) { if (obj == null) { return null; } @@ -71,6 +71,6 @@ public static com.azure.search.documents.indexes.implementation.models.PathHiera return pathHierarchyTokenizerV2; } - private PathHierarchyTokenizerV2Converter() { + private PathHierarchyTokenizerConverter() { } } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SearchIndexConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SearchIndexConverter.java index d1cbc1170972..147c7529627e 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SearchIndexConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SearchIndexConverter.java @@ -11,8 +11,8 @@ import com.azure.search.documents.indexes.models.SearchField; import com.azure.search.documents.indexes.models.SearchIndex; import com.azure.search.documents.indexes.models.SearchResourceEncryptionKey; -import com.azure.search.documents.indexes.models.Similarity; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SimilarityAlgorithm; +import com.azure.search.documents.indexes.models.SearchSuggester; import com.azure.search.documents.indexes.models.TokenFilter; import java.util.List; @@ -38,9 +38,9 @@ public static SearchIndex map(com.azure.search.documents.indexes.implementation. } if (obj.getSuggesters() != null) { - List suggesters = + List searchSuggesters = obj.getSuggesters().stream().map(SuggesterConverter::map).collect(Collectors.toList()); - searchIndex.setSuggesters(suggesters); + searchIndex.setSearchSuggesters(searchSuggesters); } if (obj.getCharFilters() != null) { @@ -71,8 +71,8 @@ public static SearchIndex map(com.azure.search.documents.indexes.implementation. } if (obj.getSimilarity() != null) { - Similarity similarity = SimilarityConverter.map(obj.getSimilarity()); - searchIndex.setSimilarity(similarity); + SimilarityAlgorithm similarityAlgorithm = SimilarityConverter.map(obj.getSimilarity()); + searchIndex.setSimilarityAlgorithm(similarityAlgorithm); } String name = obj.getName(); @@ -116,9 +116,9 @@ public static com.azure.search.documents.indexes.implementation.models.SearchInd searchIndex.setTokenizers(tokenizers); } - if (obj.getSuggesters() != null) { + if (obj.getSearchSuggesters() != null) { List suggesters = - obj.getSuggesters().stream().map(SuggesterConverter::map).collect(Collectors.toList()); + obj.getSearchSuggesters().stream().map(SuggesterConverter::map).collect(Collectors.toList()); searchIndex.setSuggesters(suggesters); } @@ -149,9 +149,9 @@ public static com.azure.search.documents.indexes.implementation.models.SearchInd searchIndex.setAnalyzers(analyzers); } - if (obj.getSimilarity() != null) { + if (obj.getSimilarityAlgorithm() != null) { com.azure.search.documents.indexes.implementation.models.Similarity similarity = - SimilarityConverter.map(obj.getSimilarity()); + SimilarityConverter.map(obj.getSimilarityAlgorithm()); searchIndex.setSimilarity(similarity); } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SimilarityConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SimilarityConverter.java index 46c18cb8b1f0..ae37a7acb6d6 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SimilarityConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SimilarityConverter.java @@ -3,27 +3,27 @@ package com.azure.search.documents.implementation.converters; -import com.azure.search.documents.indexes.models.Similarity; +import com.azure.search.documents.indexes.models.SimilarityAlgorithm; /** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.Similarity} and {@link Similarity}. + * A converter between {@link com.azure.search.documents.indexes.implementation.models.Similarity} and {@link SimilarityAlgorithm}. */ public final class SimilarityConverter { /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.Similarity} to {@link Similarity}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.Similarity} to {@link SimilarityAlgorithm}. */ - public static Similarity map(com.azure.search.documents.indexes.implementation.models.Similarity obj) { + public static SimilarityAlgorithm map(com.azure.search.documents.indexes.implementation.models.Similarity obj) { if (obj == null) { return null; } - Similarity similarity = new Similarity(); - return similarity; + SimilarityAlgorithm similarityAlgorithm = new SimilarityAlgorithm(); + return similarityAlgorithm; } /** - * Maps from {@link Similarity} to {@link com.azure.search.documents.indexes.implementation.models.Similarity}. + * Maps from {@link SimilarityAlgorithm} to {@link com.azure.search.documents.indexes.implementation.models.Similarity}. */ - public static com.azure.search.documents.indexes.implementation.models.Similarity map(Similarity obj) { + public static com.azure.search.documents.indexes.implementation.models.Similarity map(SimilarityAlgorithm obj) { if (obj == null) { return null; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SuggesterConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SuggesterConverter.java index 13af551b02ec..2da3eaa6f357 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SuggesterConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/SuggesterConverter.java @@ -3,40 +3,40 @@ package com.azure.search.documents.implementation.converters; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SearchSuggester; import java.util.ArrayList; import java.util.List; /** - * A converter between {@link com.azure.search.documents.indexes.implementation.models.Suggester} and {@link Suggester}. + * A converter between {@link com.azure.search.documents.indexes.implementation.models.Suggester} and {@link SearchSuggester}. */ public final class SuggesterConverter { private static final String SEARCH_MODE = "analyzingInfixMatching"; /** - * Maps from {@link com.azure.search.documents.indexes.implementation.models.Suggester} to {@link Suggester}. + * Maps from {@link com.azure.search.documents.indexes.implementation.models.Suggester} to {@link SearchSuggester}. */ - public static Suggester map(com.azure.search.documents.indexes.implementation.models.Suggester obj) { + public static SearchSuggester map(com.azure.search.documents.indexes.implementation.models.Suggester obj) { if (obj == null) { return null; } - Suggester suggester = new Suggester(); + SearchSuggester searchSuggester = new SearchSuggester(); if (obj.getSourceFields() != null) { List sourceFields = new ArrayList<>(obj.getSourceFields()); - suggester.setSourceFields(sourceFields); + searchSuggester.setSourceFields(sourceFields); } String name = obj.getName(); - suggester.setName(name); + searchSuggester.setName(name); - return suggester; + return searchSuggester; } /** - * Maps from {@link Suggester} to {@link com.azure.search.documents.indexes.implementation.models.Suggester}. + * Maps from {@link SearchSuggester} to {@link com.azure.search.documents.indexes.implementation.models.Suggester}. */ - public static com.azure.search.documents.indexes.implementation.models.Suggester map(Suggester obj) { + public static com.azure.search.documents.indexes.implementation.models.Suggester map(SearchSuggester obj) { if (obj == null) { return null; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java index 57a85033f001..e302b285e73b 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java @@ -9,7 +9,6 @@ import com.azure.search.documents.indexes.implementation.models.CommonGramTokenFilter; import com.azure.search.documents.indexes.implementation.models.DictionaryDecompounderTokenFilter; import com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter; -import com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2; import com.azure.search.documents.indexes.implementation.models.ElisionTokenFilter; import com.azure.search.documents.indexes.implementation.models.KeepTokenFilter; import com.azure.search.documents.indexes.implementation.models.KeywordMarkerTokenFilter; diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java index 5149d1215862..c5453429eee0 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java @@ -16,7 +16,7 @@ * of an input token. This token filter is implemented using Apache Lucene. */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.EdgeNGramTokenFilter") +@JsonTypeName("#Microsoft.Azure.Search.EdgeNGramTokenFilterV2") @Fluent public final class EdgeNGramTokenFilter extends TokenFilter { /* diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25Similarity.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25SimilarityAlgorithm.java similarity index 94% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25Similarity.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25SimilarityAlgorithm.java index c355a7b11873..01c9efe73c42 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25Similarity.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/BM25SimilarityAlgorithm.java @@ -17,7 +17,7 @@ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") @JsonTypeName("#Microsoft.Azure.Search.BM25Similarity") @Fluent -public final class BM25Similarity extends Similarity { +public final class BM25SimilarityAlgorithm extends SimilarityAlgorithm { /* * This property controls the scaling function between the term frequency * of each matching terms and the final relevance score of a document-query @@ -57,7 +57,7 @@ public Double getK1() { * @param k1 the k1 value to set. * @return the BM25Similarity object itself. */ - public BM25Similarity setK1(Double k1) { + public BM25SimilarityAlgorithm setK1(Double k1) { this.k1 = k1; return this; } @@ -83,7 +83,7 @@ public Double getB() { * @param b the b value to set. * @return the BM25Similarity object itself. */ - public BM25Similarity setB(Double b) { + public BM25SimilarityAlgorithm setB(Double b) { this.b = b; return this; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarity.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarityAlgorithm.java similarity index 90% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarity.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarityAlgorithm.java index 3f7fda38d89b..14c2a8c4b258 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarity.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/ClassicSimilarityAlgorithm.java @@ -16,5 +16,5 @@ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") @JsonTypeName("#Microsoft.Azure.Search.ClassicSimilarity") @Fluent -public final class ClassicSimilarity extends Similarity { +public final class ClassicSimilarityAlgorithm extends SimilarityAlgorithm { } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/EdgeNGramTokenFilterV2.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/EdgeNGramTokenFilterV2.java deleted file mode 100644 index d2d1681bb883..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/EdgeNGramTokenFilterV2.java +++ /dev/null @@ -1,106 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.indexes.models; - -import com.azure.core.annotation.Fluent; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.fasterxml.jackson.annotation.JsonTypeInfo; -import com.fasterxml.jackson.annotation.JsonTypeName; - -/** - * Generates n-grams of the given size(s) starting from the front or the back - * of an input token. This token filter is implemented using Apache Lucene. - */ -@JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.EdgeNGramTokenFilterV2") -@Fluent -public final class EdgeNGramTokenFilterV2 extends TokenFilter { - /* - * The minimum n-gram length. Default is 1. Maximum is 300. Must be less - * than the value of maxGram. - */ - @JsonProperty(value = "minGram") - private Integer minGram; - - /* - * The maximum n-gram length. Default is 2. Maximum is 300. - */ - @JsonProperty(value = "maxGram") - private Integer maxGram; - - /* - * Specifies which side of the input the n-gram should be generated from. - * Default is "front". Possible values include: 'Front', 'Back' - */ - @JsonProperty(value = "side") - private EdgeNGramTokenFilterSide side; - - /** - * Get the minGram property: The minimum n-gram length. Default is 1. - * Maximum is 300. Must be less than the value of maxGram. - * - * @return the minGram value. - */ - public Integer getMinGram() { - return this.minGram; - } - - /** - * Set the minGram property: The minimum n-gram length. Default is 1. - * Maximum is 300. Must be less than the value of maxGram. - * - * @param minGram the minGram value to set. - * @return the EdgeNGramTokenFilterV2 object itself. - */ - public EdgeNGramTokenFilterV2 setMinGram(Integer minGram) { - this.minGram = minGram; - return this; - } - - /** - * Get the maxGram property: The maximum n-gram length. Default is 2. - * Maximum is 300. - * - * @return the maxGram value. - */ - public Integer getMaxGram() { - return this.maxGram; - } - - /** - * Set the maxGram property: The maximum n-gram length. Default is 2. - * Maximum is 300. - * - * @param maxGram the maxGram value to set. - * @return the EdgeNGramTokenFilterV2 object itself. - */ - public EdgeNGramTokenFilterV2 setMaxGram(Integer maxGram) { - this.maxGram = maxGram; - return this; - } - - /** - * Get the side property: Specifies which side of the input the n-gram - * should be generated from. Default is "front". Possible values include: - * 'Front', 'Back'. - * - * @return the side value. - */ - public EdgeNGramTokenFilterSide getSide() { - return this.side; - } - - /** - * Set the side property: Specifies which side of the input the n-gram - * should be generated from. Default is "front". Possible values include: - * 'Front', 'Back'. - * - * @param side the side value to set. - * @return the EdgeNGramTokenFilterV2 object itself. - */ - public EdgeNGramTokenFilterV2 setSide(EdgeNGramTokenFilterSide side) { - this.side = side; - return this; - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizer.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizer.java index f5fda9e3e318..0b2e572f6f96 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizer.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizer.java @@ -13,34 +13,38 @@ * using Apache Lucene. */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.KeywordTokenizer") +@JsonTypeName("#Microsoft.Azure.Search.KeywordTokenizerV2") @Fluent public final class KeywordTokenizer extends LexicalTokenizer { /* - * The read buffer size in bytes. Default is 256. + * The maximum token length. Default is 256. Tokens longer than the maximum + * length are split. The maximum token length that can be used is 300 + * characters. */ - @JsonProperty(value = "bufferSize") - private Integer bufferSize; + @JsonProperty(value = "maxTokenLength") + private Integer maxTokenLength; /** - * Get the bufferSize property: The read buffer size in bytes. Default is - * 256. + * Get the maxTokenLength property: The maximum token length. Default is + * 256. Tokens longer than the maximum length are split. The maximum token + * length that can be used is 300 characters. * - * @return the bufferSize value. + * @return the maxTokenLength value. */ - public Integer getBufferSize() { - return this.bufferSize; + public Integer getMaxTokenLength() { + return this.maxTokenLength; } /** - * Set the bufferSize property: The read buffer size in bytes. Default is - * 256. + * Set the maxTokenLength property: The maximum token length. Default is + * 256. Tokens longer than the maximum length are split. The maximum token + * length that can be used is 300 characters. * - * @param bufferSize the bufferSize value to set. - * @return the KeywordTokenizer object itself. + * @param maxTokenLength the maxTokenLength value to set. + * @return the KeywordTokenizerV2 object itself. */ - public KeywordTokenizer setBufferSize(Integer bufferSize) { - this.bufferSize = bufferSize; + public KeywordTokenizer setMaxTokenLength(Integer maxTokenLength) { + this.maxTokenLength = maxTokenLength; return this; } } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizerV2.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizerV2.java deleted file mode 100644 index 92c5bf47c0ce..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/KeywordTokenizerV2.java +++ /dev/null @@ -1,50 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.indexes.models; - -import com.azure.core.annotation.Fluent; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.fasterxml.jackson.annotation.JsonTypeInfo; -import com.fasterxml.jackson.annotation.JsonTypeName; - -/** - * Emits the entire input as a single token. This tokenizer is implemented - * using Apache Lucene. - */ -@JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.KeywordTokenizerV2") -@Fluent -public final class KeywordTokenizerV2 extends LexicalTokenizer { - /* - * The maximum token length. Default is 256. Tokens longer than the maximum - * length are split. The maximum token length that can be used is 300 - * characters. - */ - @JsonProperty(value = "maxTokenLength") - private Integer maxTokenLength; - - /** - * Get the maxTokenLength property: The maximum token length. Default is - * 256. Tokens longer than the maximum length are split. The maximum token - * length that can be used is 300 characters. - * - * @return the maxTokenLength value. - */ - public Integer getMaxTokenLength() { - return this.maxTokenLength; - } - - /** - * Set the maxTokenLength property: The maximum token length. Default is - * 256. Tokens longer than the maximum length are split. The maximum token - * length that can be used is 300 characters. - * - * @param maxTokenLength the maxTokenLength value to set. - * @return the KeywordTokenizerV2 object itself. - */ - public KeywordTokenizerV2 setMaxTokenLength(Integer maxTokenLength) { - this.maxTokenLength = maxTokenLength; - return this; - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java index 0df97a4f0e70..7ada0ab24da4 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java @@ -19,14 +19,14 @@ @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.ClassicTokenizer", value = ClassicTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.EdgeNGramTokenizer", value = EdgeNGramTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.KeywordTokenizer", value = KeywordTokenizer.class), - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.KeywordTokenizerV2", value = KeywordTokenizerV2.class), + @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.KeywordTokenizerV2", value = KeywordTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.MicrosoftLanguageTokenizer", value = MicrosoftLanguageTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.MicrosoftLanguageStemmingTokenizer", value = MicrosoftLanguageStemmingTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.NGramTokenizer", value = NGramTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.PathHierarchyTokenizerV2", - value = PathHierarchyTokenizerV2.class), + value = PathHierarchyTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.PatternTokenizer", value = PatternTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.StandardTokenizer", value = LuceneStandardTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.StandardTokenizerV2", value = LuceneStandardTokenizerV2.class), diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LuceneStandardTokenizerV2.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LuceneStandardTokenizerV2.java deleted file mode 100644 index fc1218bf418e..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LuceneStandardTokenizerV2.java +++ /dev/null @@ -1,50 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.indexes.models; - -import com.azure.core.annotation.Fluent; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.fasterxml.jackson.annotation.JsonTypeInfo; -import com.fasterxml.jackson.annotation.JsonTypeName; - -/** - * Breaks text following the Unicode Text Segmentation rules. This tokenizer is - * implemented using Apache Lucene. - */ -@JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.StandardTokenizerV2") -@Fluent -public final class LuceneStandardTokenizerV2 extends LexicalTokenizer { - /* - * The maximum token length. Default is 255. Tokens longer than the maximum - * length are split. The maximum token length that can be used is 300 - * characters. - */ - @JsonProperty(value = "maxTokenLength") - private Integer maxTokenLength; - - /** - * Get the maxTokenLength property: The maximum token length. Default is - * 255. Tokens longer than the maximum length are split. The maximum token - * length that can be used is 300 characters. - * - * @return the maxTokenLength value. - */ - public Integer getMaxTokenLength() { - return this.maxTokenLength; - } - - /** - * Set the maxTokenLength property: The maximum token length. Default is - * 255. Tokens longer than the maximum length are split. The maximum token - * length that can be used is 300 characters. - * - * @param maxTokenLength the maxTokenLength value to set. - * @return the LuceneStandardTokenizerV2 object itself. - */ - public LuceneStandardTokenizerV2 setMaxTokenLength(Integer maxTokenLength) { - this.maxTokenLength = maxTokenLength; - return this; - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/NGramTokenFilterV2.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/NGramTokenFilterV2.java deleted file mode 100644 index 690427f22298..000000000000 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/NGramTokenFilterV2.java +++ /dev/null @@ -1,75 +0,0 @@ -// Copyright (c) Microsoft Corporation. All rights reserved. -// Licensed under the MIT License. - -package com.azure.search.documents.indexes.models; - -import com.azure.core.annotation.Fluent; -import com.fasterxml.jackson.annotation.JsonProperty; -import com.fasterxml.jackson.annotation.JsonTypeInfo; -import com.fasterxml.jackson.annotation.JsonTypeName; - -/** - * Generates n-grams of the given size(s). This token filter is implemented - * using Apache Lucene. - */ -@JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.NGramTokenFilterV2") -@Fluent -public final class NGramTokenFilterV2 extends TokenFilter { - /* - * The minimum n-gram length. Default is 1. Maximum is 300. Must be less - * than the value of maxGram. - */ - @JsonProperty(value = "minGram") - private Integer minGram; - - /* - * The maximum n-gram length. Default is 2. Maximum is 300. - */ - @JsonProperty(value = "maxGram") - private Integer maxGram; - - /** - * Get the minGram property: The minimum n-gram length. Default is 1. - * Maximum is 300. Must be less than the value of maxGram. - * - * @return the minGram value. - */ - public Integer getMinGram() { - return this.minGram; - } - - /** - * Set the minGram property: The minimum n-gram length. Default is 1. - * Maximum is 300. Must be less than the value of maxGram. - * - * @param minGram the minGram value to set. - * @return the NGramTokenFilterV2 object itself. - */ - public NGramTokenFilterV2 setMinGram(Integer minGram) { - this.minGram = minGram; - return this; - } - - /** - * Get the maxGram property: The maximum n-gram length. Default is 2. - * Maximum is 300. - * - * @return the maxGram value. - */ - public Integer getMaxGram() { - return this.maxGram; - } - - /** - * Set the maxGram property: The maximum n-gram length. Default is 2. - * Maximum is 300. - * - * @param maxGram the maxGram value to set. - * @return the NGramTokenFilterV2 object itself. - */ - public NGramTokenFilterV2 setMaxGram(Integer maxGram) { - this.maxGram = maxGram; - return this; - } -} diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizerV2.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizer.java similarity index 90% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizerV2.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizer.java index 7e3e3bccb541..d74650a2add5 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizerV2.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/PathHierarchyTokenizer.java @@ -15,7 +15,7 @@ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") @JsonTypeName("#Microsoft.Azure.Search.PathHierarchyTokenizerV2") @Fluent -public final class PathHierarchyTokenizerV2 extends LexicalTokenizer { +public final class PathHierarchyTokenizer extends LexicalTokenizer { /* * The delimiter character to use. Default is "/". */ @@ -64,7 +64,7 @@ public String getDelimiter() { * @param delimiter the delimiter value to set. * @return the PathHierarchyTokenizerV2 object itself. */ - public PathHierarchyTokenizerV2 setDelimiter(String delimiter) { + public PathHierarchyTokenizer setDelimiter(String delimiter) { this.delimiter = delimiter; return this; } @@ -86,7 +86,7 @@ public String getReplacement() { * @param replacement the replacement value to set. * @return the PathHierarchyTokenizerV2 object itself. */ - public PathHierarchyTokenizerV2 setReplacement(String replacement) { + public PathHierarchyTokenizer setReplacement(String replacement) { this.replacement = replacement; return this; } @@ -108,7 +108,7 @@ public Integer getMaxTokenLength() { * @param maxTokenLength the maxTokenLength value to set. * @return the PathHierarchyTokenizerV2 object itself. */ - public PathHierarchyTokenizerV2 setMaxTokenLength(Integer maxTokenLength) { + public PathHierarchyTokenizer setMaxTokenLength(Integer maxTokenLength) { this.maxTokenLength = maxTokenLength; return this; } @@ -130,7 +130,7 @@ public Boolean isReverseTokenOrder() { * @param reverseTokenOrder the reverseTokenOrder value to set. * @return the PathHierarchyTokenizerV2 object itself. */ - public PathHierarchyTokenizerV2 setReverseTokenOrder(Boolean reverseTokenOrder) { + public PathHierarchyTokenizer setReverseTokenOrder(Boolean reverseTokenOrder) { this.reverseTokenOrder = reverseTokenOrder; return this; } @@ -152,7 +152,7 @@ public Integer getNumberOfTokensToSkip() { * @param numberOfTokensToSkip the numberOfTokensToSkip value to set. * @return the PathHierarchyTokenizerV2 object itself. */ - public PathHierarchyTokenizerV2 setNumberOfTokensToSkip(Integer numberOfTokensToSkip) { + public PathHierarchyTokenizer setNumberOfTokensToSkip(Integer numberOfTokensToSkip) { this.numberOfTokensToSkip = numberOfTokensToSkip; return this; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchIndex.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchIndex.java index de82fa636e79..8cd3d95b083c 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchIndex.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchIndex.java @@ -49,7 +49,7 @@ public final class SearchIndex { * The suggesters for the index. */ @JsonProperty(value = "suggesters") - private List suggesters; + private List searchSuggesters; /* * The analyzers for the index. @@ -97,7 +97,7 @@ public final class SearchIndex { * indexes. If null, the ClassicSimilarity algorithm is used. */ @JsonProperty(value = "similarity") - private Similarity similarity; + private SimilarityAlgorithm similarityAlgorithm; /* * The ETag of the index. @@ -218,18 +218,18 @@ public SearchIndex setCorsOptions(CorsOptions corsOptions) { * * @return the suggesters value. */ - public List getSuggesters() { - return this.suggesters; + public List getSearchSuggesters() { + return this.searchSuggesters; } /** * Set the suggesters property: The suggesters for the index. * - * @param suggesters the suggesters value to set. + * @param searchSuggesters the suggesters value to set. * @return the SearchIndex object itself. */ - public SearchIndex setSuggesters(List suggesters) { - this.suggesters = suggesters; + public SearchIndex setSearchSuggesters(List searchSuggesters) { + this.searchSuggesters = searchSuggesters; return this; } @@ -362,8 +362,8 @@ public SearchIndex setEncryptionKey(SearchResourceEncryptionKey encryptionKey) { * * @return the similarity value. */ - public Similarity getSimilarity() { - return this.similarity; + public SimilarityAlgorithm getSimilarityAlgorithm() { + return this.similarityAlgorithm; } /** @@ -373,11 +373,11 @@ public Similarity getSimilarity() { * cannot be modified on existing indexes. If null, the ClassicSimilarity * algorithm is used. * - * @param similarity the similarity value to set. + * @param similarityAlgorithm the similarity value to set. * @return the SearchIndex object itself. */ - public SearchIndex setSimilarity(Similarity similarity) { - this.similarity = similarity; + public SearchIndex setSimilarityAlgorithm(SimilarityAlgorithm similarityAlgorithm) { + this.similarityAlgorithm = similarityAlgorithm; return this; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Suggester.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchSuggester.java similarity index 92% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Suggester.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchSuggester.java index b8d3f224aad7..1b5cca6a54cf 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Suggester.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SearchSuggester.java @@ -11,7 +11,7 @@ * Defines how the Suggest API should apply to a group of fields in the index. */ @Fluent -public final class Suggester { +public final class SearchSuggester { /* * A value indicating the capabilities of the suggester. */ @@ -46,7 +46,7 @@ public String getName() { * @param name the name value to set. * @return the Suggester object itself. */ - public Suggester setName(String name) { + public SearchSuggester setName(String name) { this.name = name; return this; } @@ -68,7 +68,7 @@ public List getSourceFields() { * @param sourceFields the sourceFields value to set. * @return the Suggester object itself. */ - public Suggester setSourceFields(List sourceFields) { + public SearchSuggester setSourceFields(List sourceFields) { this.sourceFields = sourceFields; return this; } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Similarity.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SimilarityAlgorithm.java similarity index 83% rename from sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Similarity.java rename to sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SimilarityAlgorithm.java index 85c4288e0091..265ed0e8a515 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/Similarity.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/SimilarityAlgorithm.java @@ -15,12 +15,12 @@ * to rank the search results. */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type", - defaultImpl = Similarity.class) + defaultImpl = SimilarityAlgorithm.class) @JsonTypeName("Similarity") @JsonSubTypes({ - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.ClassicSimilarity", value = ClassicSimilarity.class), - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.BM25Similarity", value = BM25Similarity.class) + @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.ClassicSimilarity", value = ClassicSimilarityAlgorithm.class), + @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.BM25Similarity", value = BM25SimilarityAlgorithm.class) }) @Fluent -public class Similarity { +public class SimilarityAlgorithm { } diff --git a/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/IndexAndServiceStatisticsExample.java b/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/IndexAndServiceStatisticsExample.java index 6909acc871d6..15054554fc15 100644 --- a/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/IndexAndServiceStatisticsExample.java +++ b/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/IndexAndServiceStatisticsExample.java @@ -22,7 +22,7 @@ import com.azure.search.documents.indexes.models.SearchFieldDataType; import com.azure.search.documents.indexes.models.SearchIndex; import com.azure.search.documents.indexes.models.ServiceStatistics; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SearchSuggester; import com.azure.search.documents.indexes.models.TagScoringFunction; import com.azure.search.documents.indexes.models.TagScoringParameters; import com.azure.search.documents.indexes.models.TextWeights; @@ -412,7 +412,7 @@ private static SearchIndex createTestIndex() { .setCorsOptions(new CorsOptions() .setAllowedOrigins("http://tempuri.org", "http://localhost:80") .setMaxAgeInSeconds(60L)) - .setSuggesters(Collections.singletonList(new Suggester() + .setSearchSuggesters(Collections.singletonList(new SearchSuggester() .setName("FancySuggester") .setSourceFields(Collections.singletonList("HotelName")))); } diff --git a/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/indexes/LifecycleSetupExample.java b/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/indexes/LifecycleSetupExample.java index 1fb4baf70a75..17b2e661a577 100644 --- a/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/indexes/LifecycleSetupExample.java +++ b/sdk/search/azure-search-documents/src/samples/java/com/azure/search/documents/indexes/LifecycleSetupExample.java @@ -21,7 +21,7 @@ import com.azure.search.documents.indexes.models.SearchIndexerDataSourceType; import com.azure.search.documents.indexes.models.SearchIndexerSkill; import com.azure.search.documents.indexes.models.SearchIndexerSkillset; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SearchSuggester; import java.time.Duration; import java.util.Arrays; @@ -187,7 +187,7 @@ private static SearchIndex createIndex(SearchIndexClient client) { .setAnalyzer(LexicalAnalyzerName.EN_MICROSOFT))); // Set Suggester - index.setSuggesters(Collections.singletonList(new Suggester() + index.setSearchSuggesters(Collections.singletonList(new SearchSuggester() .setName(SUGGESTER_NAME) .setSourceFields(Collections.singletonList("Tags")))); diff --git a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/SearchTestBase.java b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/SearchTestBase.java index c89b376e54c2..f182f61498e2 100644 --- a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/SearchTestBase.java +++ b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/SearchTestBase.java @@ -30,7 +30,7 @@ import com.azure.search.documents.indexes.models.SearchIndex; import com.azure.search.documents.indexes.models.SearchIndexerDataSource; import com.azure.search.documents.indexes.models.SoftDeleteColumnDeletionDetectionPolicy; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SearchSuggester; import com.azure.search.documents.indexes.models.TagScoringFunction; import com.azure.search.documents.indexes.models.TagScoringParameters; import com.azure.search.documents.indexes.models.TextWeights; @@ -451,7 +451,7 @@ protected SearchIndex createTestIndex() { .setCorsOptions(new CorsOptions() .setAllowedOrigins("http://tempuri.org", "http://localhost:80") .setMaxAgeInSeconds(60L)) - .setSuggesters(Collections.singletonList(new Suggester() + .setSearchSuggesters(Collections.singletonList(new SearchSuggester() .setName("FancySuggester") .setSourceFields(Collections.singletonList("HotelName")))); } diff --git a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java index 2a46ef441096..ec2e824d23ac 100644 --- a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java +++ b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java @@ -19,12 +19,11 @@ import com.azure.search.documents.indexes.models.CustomAnalyzer; import com.azure.search.documents.indexes.models.DictionaryDecompounderTokenFilter; import com.azure.search.documents.indexes.models.EdgeNGramTokenFilterSide; -import com.azure.search.documents.indexes.models.EdgeNGramTokenFilterV2; import com.azure.search.documents.indexes.models.EdgeNGramTokenizer; import com.azure.search.documents.indexes.models.ElisionTokenFilter; import com.azure.search.documents.indexes.models.KeepTokenFilter; import com.azure.search.documents.indexes.models.KeywordMarkerTokenFilter; -import com.azure.search.documents.indexes.models.KeywordTokenizerV2; +import com.azure.search.documents.indexes.models.KeywordTokenizer; import com.azure.search.documents.indexes.models.LengthTokenFilter; import com.azure.search.documents.indexes.models.LexicalAnalyzer; import com.azure.search.documents.indexes.models.LexicalAnalyzerName; @@ -32,15 +31,13 @@ import com.azure.search.documents.indexes.models.LexicalTokenizerName; import com.azure.search.documents.indexes.models.LimitTokenFilter; import com.azure.search.documents.indexes.models.LuceneStandardAnalyzer; -import com.azure.search.documents.indexes.models.LuceneStandardTokenizerV2; import com.azure.search.documents.indexes.models.MappingCharFilter; import com.azure.search.documents.indexes.models.MicrosoftLanguageStemmingTokenizer; import com.azure.search.documents.indexes.models.MicrosoftLanguageTokenizer; import com.azure.search.documents.indexes.models.MicrosoftStemmingTokenizerLanguage; import com.azure.search.documents.indexes.models.MicrosoftTokenizerLanguage; -import com.azure.search.documents.indexes.models.NGramTokenFilterV2; import com.azure.search.documents.indexes.models.NGramTokenizer; -import com.azure.search.documents.indexes.models.PathHierarchyTokenizerV2; +import com.azure.search.documents.indexes.models.PathHierarchyTokenizer; import com.azure.search.documents.indexes.models.PatternAnalyzer; import com.azure.search.documents.indexes.models.PatternCaptureTokenFilter; import com.azure.search.documents.indexes.models.PatternReplaceCharFilter; @@ -952,7 +949,7 @@ SearchIndex prepareIndexWithAllAnalysisComponentTypes() { new ClassicTokenizer() .setMaxTokenLength(100) .setName(generateName()), - new KeywordTokenizerV2() + new KeywordTokenizer() .setMaxTokenLength(100) .setName(generateName()), new MicrosoftLanguageStemmingTokenizer() @@ -965,7 +962,7 @@ SearchIndex prepareIndexWithAllAnalysisComponentTypes() { .setIsSearchTokenizer(true) .setLanguage(MicrosoftTokenizerLanguage.THAI) .setName(generateName()), - new PathHierarchyTokenizerV2() + new PathHierarchyTokenizer() .setDelimiter(":") .setReplacement("_") .setMaxTokenLength(300) @@ -1143,13 +1140,13 @@ SearchIndex createIndexWithSpecialDefaults() { .setName(generateSimpleName(i++)), new ClassicTokenizer() .setName(generateSimpleName(i++)), - new KeywordTokenizerV2() + new KeywordTokenizer() .setName(generateSimpleName(i++)), new MicrosoftLanguageStemmingTokenizer() .setName(generateSimpleName(i++)), new MicrosoftLanguageTokenizer() .setName(generateSimpleName(i++)), - new PathHierarchyTokenizerV2() + new PathHierarchyTokenizer() .setName(generateSimpleName(i++)), new PatternTokenizer() .setName(generateSimpleName(i++)), @@ -1215,7 +1212,7 @@ SearchIndex createExpectedIndexWithSpecialDefaults(SearchIndex index) { new ClassicTokenizer() .setMaxTokenLength(255) .setName(generateSimpleName(i++)), - new KeywordTokenizerV2() + new KeywordTokenizer() .setMaxTokenLength(256) .setName(generateSimpleName(i++)), new MicrosoftLanguageStemmingTokenizer() @@ -1228,7 +1225,7 @@ SearchIndex createExpectedIndexWithSpecialDefaults(SearchIndex index) { .setIsSearchTokenizer(false) .setLanguage(MicrosoftTokenizerLanguage.ENGLISH) .setName(generateSimpleName(i++)), - new PathHierarchyTokenizerV2() + new PathHierarchyTokenizer() .setDelimiter("/") .setReplacement("/") .setMaxTokenLength(300) diff --git a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/IndexManagementSyncTests.java b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/IndexManagementSyncTests.java index 91e7cb7b12d7..2cc7bfd66849 100644 --- a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/IndexManagementSyncTests.java +++ b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/IndexManagementSyncTests.java @@ -19,7 +19,7 @@ import com.azure.search.documents.indexes.models.SearchField; import com.azure.search.documents.indexes.models.SearchFieldDataType; import com.azure.search.documents.indexes.models.SearchIndex; -import com.azure.search.documents.indexes.models.Suggester; +import com.azure.search.documents.indexes.models.SearchSuggester; import com.azure.search.documents.indexes.models.SynonymMap; import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.Test; @@ -412,7 +412,7 @@ public void canUpdateSuggesterWithNewIndexFields() { new SearchField() .setName("HotelRewards") .setType(SearchFieldDataType.STRING))); - existingIndex.setSuggesters(Collections.singletonList(new Suggester() + existingIndex.setSearchSuggesters(Collections.singletonList(new SearchSuggester() .setName("Suggestion") .setSourceFields(Arrays.asList("HotelAmenities", "HotelRewards")) )); @@ -430,7 +430,7 @@ public void createOrUpdateIndexThrowsWhenUpdatingSuggesterWithExistingIndexField SearchIndex existingIndex = client.getIndex(index.getName()); String existingFieldName = "Category"; - existingIndex.setSuggesters(Collections.singletonList(new Suggester() + existingIndex.setSearchSuggesters(Collections.singletonList(new SearchSuggester() .setName("Suggestion") .setSourceFields(Collections.singletonList(existingFieldName)) )); From e11404deaaadad0c393484c773dded017aef6708 Mon Sep 17 00:00:00 2001 From: Sima Zhu Date: Fri, 29 May 2020 11:17:53 -0700 Subject: [PATCH 2/4] Fixed compile issues --- .../converters/LexicalTokenizerConverter.java | 5 +--- .../converters/TokenFilterConverter.java | 11 +++----- .../indexes/models/LexicalTokenizer.java | 1 - .../documents/indexes/models/TokenFilter.java | 2 -- .../indexes/CustomAnalyzerSyncTests.java | 27 ++++++++++--------- 5 files changed, 19 insertions(+), 27 deletions(-) diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java index c5a46773e22e..2bb239c17046 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/LexicalTokenizerConverter.java @@ -50,7 +50,7 @@ public static LexicalTokenizer map(com.azure.search.documents.indexes.implementa return KeywordTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.KeywordTokenizer) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2) { - return LuceneStandardTokenizerV2Converter.map((com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2) obj); + return LuceneStandardTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.LuceneStandardTokenizerV2) obj); } if (obj instanceof com.azure.search.documents.indexes.implementation.models.UaxUrlEmailTokenizer) { return UaxUrlEmailTokenizerConverter.map((com.azure.search.documents.indexes.implementation.models.UaxUrlEmailTokenizer) obj); @@ -94,9 +94,6 @@ public static com.azure.search.documents.indexes.implementation.models.LexicalTo if (obj instanceof MicrosoftLanguageTokenizer) { return MicrosoftLanguageTokenizerConverter.map((MicrosoftLanguageTokenizer) obj); } - if (obj instanceof LuceneStandardTokenizerV2) { - return LuceneStandardTokenizerV2Converter.map((LuceneStandardTokenizerV2) obj); - } if (obj instanceof LuceneStandardTokenizer) { return LuceneStandardTokenizerConverter.map((LuceneStandardTokenizer) obj); } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java index e302b285e73b..b57bf8ed48e9 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/implementation/converters/TokenFilterConverter.java @@ -9,6 +9,7 @@ import com.azure.search.documents.indexes.implementation.models.CommonGramTokenFilter; import com.azure.search.documents.indexes.implementation.models.DictionaryDecompounderTokenFilter; import com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilter; +import com.azure.search.documents.indexes.implementation.models.EdgeNGramTokenFilterV2; import com.azure.search.documents.indexes.implementation.models.ElisionTokenFilter; import com.azure.search.documents.indexes.implementation.models.KeepTokenFilter; import com.azure.search.documents.indexes.implementation.models.KeywordMarkerTokenFilter; @@ -72,13 +73,13 @@ public static TokenFilter map(com.azure.search.documents.indexes.implementation. return CjkBigramTokenFilterConverter.map((CjkBigramTokenFilter) obj); } if (obj instanceof EdgeNGramTokenFilterV2) { - return EdgeNGramTokenFilterV2Converter.map((EdgeNGramTokenFilterV2) obj); + return EdgeNGramTokenFilterConverter.map((EdgeNGramTokenFilterV2) obj); } if (obj instanceof PatternCaptureTokenFilter) { return PatternCaptureTokenFilterConverter.map((PatternCaptureTokenFilter) obj); } if (obj instanceof NGramTokenFilterV2) { - return NGramTokenFilterV2Converter.map((NGramTokenFilterV2) obj); + return NGramTokenFilterConverter.map((NGramTokenFilterV2) obj); } if (obj instanceof PatternReplaceTokenFilter) { return PatternReplaceTokenFilterConverter.map((PatternReplaceTokenFilter) obj); @@ -164,18 +165,12 @@ public static com.azure.search.documents.indexes.implementation.models.TokenFilt if (obj instanceof com.azure.search.documents.indexes.models.EdgeNGramTokenFilter) { return EdgeNGramTokenFilterConverter.map((com.azure.search.documents.indexes.models.EdgeNGramTokenFilter) obj); } - if (obj instanceof com.azure.search.documents.indexes.models.EdgeNGramTokenFilterV2) { - return EdgeNGramTokenFilterV2Converter.map((com.azure.search.documents.indexes.models.EdgeNGramTokenFilterV2) obj); - } if (obj instanceof com.azure.search.documents.indexes.models.DictionaryDecompounderTokenFilter) { return DictionaryDecompounderTokenFilterConverter.map((com.azure.search.documents.indexes.models.DictionaryDecompounderTokenFilter) obj); } if (obj instanceof com.azure.search.documents.indexes.models.PatternReplaceTokenFilter) { return PatternReplaceTokenFilterConverter.map((com.azure.search.documents.indexes.models.PatternReplaceTokenFilter) obj); } - if (obj instanceof com.azure.search.documents.indexes.models.NGramTokenFilterV2) { - return NGramTokenFilterV2Converter.map((com.azure.search.documents.indexes.models.NGramTokenFilterV2) obj); - } if (obj instanceof com.azure.search.documents.indexes.models.StemmerTokenFilter) { return StemmerTokenFilterConverter.map((com.azure.search.documents.indexes.models.StemmerTokenFilter) obj); } diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java index 7ada0ab24da4..54f11d53f4da 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/LexicalTokenizer.java @@ -29,7 +29,6 @@ value = PathHierarchyTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.PatternTokenizer", value = PatternTokenizer.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.StandardTokenizer", value = LuceneStandardTokenizer.class), - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.StandardTokenizerV2", value = LuceneStandardTokenizerV2.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.UaxUrlEmailTokenizer", value = UaxUrlEmailTokenizer.class) }) @Fluent diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/TokenFilter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/TokenFilter.java index 11db7d99dcb6..f3ebdf4c2380 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/TokenFilter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/models/TokenFilter.java @@ -22,7 +22,6 @@ @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.DictionaryDecompounderTokenFilter", value = DictionaryDecompounderTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.EdgeNGramTokenFilter", value = EdgeNGramTokenFilter.class), - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.EdgeNGramTokenFilterV2", value = EdgeNGramTokenFilterV2.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.ElisionTokenFilter", value = ElisionTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.KeepTokenFilter", value = KeepTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.KeywordMarkerTokenFilter", @@ -30,7 +29,6 @@ @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.LengthTokenFilter", value = LengthTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.LimitTokenFilter", value = LimitTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.NGramTokenFilter", value = NGramTokenFilter.class), - @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.NGramTokenFilterV2", value = NGramTokenFilterV2.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.PatternCaptureTokenFilter", value = PatternCaptureTokenFilter.class), @JsonSubTypes.Type(name = "#Microsoft.Azure.Search.PatternReplaceTokenFilter", diff --git a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java index ec2e824d23ac..3b45221edadc 100644 --- a/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java +++ b/sdk/search/azure-search-documents/src/test/java/com/azure/search/documents/indexes/CustomAnalyzerSyncTests.java @@ -18,6 +18,7 @@ import com.azure.search.documents.indexes.models.CommonGramTokenFilter; import com.azure.search.documents.indexes.models.CustomAnalyzer; import com.azure.search.documents.indexes.models.DictionaryDecompounderTokenFilter; +import com.azure.search.documents.indexes.models.EdgeNGramTokenFilter; import com.azure.search.documents.indexes.models.EdgeNGramTokenFilterSide; import com.azure.search.documents.indexes.models.EdgeNGramTokenizer; import com.azure.search.documents.indexes.models.ElisionTokenFilter; @@ -31,11 +32,13 @@ import com.azure.search.documents.indexes.models.LexicalTokenizerName; import com.azure.search.documents.indexes.models.LimitTokenFilter; import com.azure.search.documents.indexes.models.LuceneStandardAnalyzer; +import com.azure.search.documents.indexes.models.LuceneStandardTokenizer; import com.azure.search.documents.indexes.models.MappingCharFilter; import com.azure.search.documents.indexes.models.MicrosoftLanguageStemmingTokenizer; import com.azure.search.documents.indexes.models.MicrosoftLanguageTokenizer; import com.azure.search.documents.indexes.models.MicrosoftStemmingTokenizerLanguage; import com.azure.search.documents.indexes.models.MicrosoftTokenizerLanguage; +import com.azure.search.documents.indexes.models.NGramTokenFilter; import com.azure.search.documents.indexes.models.NGramTokenizer; import com.azure.search.documents.indexes.models.PathHierarchyTokenizer; import com.azure.search.documents.indexes.models.PatternAnalyzer; @@ -49,8 +52,6 @@ import com.azure.search.documents.indexes.models.SearchField; import com.azure.search.documents.indexes.models.SearchFieldDataType; import com.azure.search.documents.indexes.models.SearchIndex; -import com.azure.search.documents.models.SearchOptions; -import com.azure.search.documents.models.SearchResult; import com.azure.search.documents.indexes.models.ShingleTokenFilter; import com.azure.search.documents.indexes.models.SnowballTokenFilter; import com.azure.search.documents.indexes.models.SnowballTokenFilterLanguage; @@ -68,6 +69,8 @@ import com.azure.search.documents.indexes.models.UaxUrlEmailTokenizer; import com.azure.search.documents.indexes.models.UniqueTokenFilter; import com.azure.search.documents.indexes.models.WordDelimiterTokenFilter; +import com.azure.search.documents.models.SearchOptions; +import com.azure.search.documents.models.SearchResult; import org.junit.jupiter.api.Assertions; import org.junit.jupiter.api.Test; @@ -728,7 +731,7 @@ List prepareIndexesWithAllAnalysisComponentOptions() { .setName(generateName())); tokenFilters.addAll( Arrays.stream(EdgeNGramTokenFilterSide.values()) - .map(s -> new EdgeNGramTokenFilterV2() + .map(s -> new EdgeNGramTokenFilter() .setMinGram(1) .setMaxGram(2) .setSide(s) @@ -974,7 +977,7 @@ SearchIndex prepareIndexWithAllAnalysisComponentTypes() { .setFlags(Collections.singletonList(RegexFlags.MULTILINE)) .setGroup(0) .setName(generateName()), - new LuceneStandardTokenizerV2() + new LuceneStandardTokenizer() .setMaxTokenLength(100) .setName(generateName()), new UaxUrlEmailTokenizer() @@ -1010,7 +1013,7 @@ SearchIndex prepareIndexWithAllAnalysisComponentTypes() { .setMaxSubwordSize(13) .setOnlyLongestMatch(true) .setName(generateName()), - new EdgeNGramTokenFilterV2() + new EdgeNGramTokenFilter() .setMinGram(2) .setMaxGram(10) .setSide(EdgeNGramTokenFilterSide.BACK) @@ -1040,7 +1043,7 @@ SearchIndex prepareIndexWithAllAnalysisComponentTypes() { .setMaxTokenCount(10) .setConsumeAllTokens(true) .setName(generateName()), - new NGramTokenFilterV2() + new NGramTokenFilter() .setMinGram(2) .setMaxGram(3) .setName(generateName()), @@ -1150,7 +1153,7 @@ SearchIndex createIndexWithSpecialDefaults() { .setName(generateSimpleName(i++)), new PatternTokenizer() .setName(generateSimpleName(i++)), - new LuceneStandardTokenizerV2() + new LuceneStandardTokenizer() .setName(generateSimpleName(i++)), new UaxUrlEmailTokenizer() .setName(generateSimpleName(i++)) @@ -1159,13 +1162,13 @@ SearchIndex createIndexWithSpecialDefaults() { new DictionaryDecompounderTokenFilter() .setWordList(Collections.singletonList("Bahnhof")) .setName(generateSimpleName(i++)), - new EdgeNGramTokenFilterV2() + new EdgeNGramTokenFilter() .setName(generateSimpleName(i++)), new LengthTokenFilter() .setName(generateSimpleName(i++)), new LimitTokenFilter() .setName(generateSimpleName(i++)), - new NGramTokenFilterV2() + new NGramTokenFilter() .setName(generateSimpleName(i++)), new PatternCaptureTokenFilter() .setPatterns(Collections.singletonList("[a-z]*")) @@ -1234,7 +1237,7 @@ SearchIndex createExpectedIndexWithSpecialDefaults(SearchIndex index) { .setPattern("\\W+") .setGroup(-1) .setName(generateSimpleName(i++)), - new LuceneStandardTokenizerV2() + new LuceneStandardTokenizer() .setMaxTokenLength(255) .setName(generateSimpleName(i++)), new UaxUrlEmailTokenizer() @@ -1248,7 +1251,7 @@ SearchIndex createExpectedIndexWithSpecialDefaults(SearchIndex index) { .setMinSubwordSize(2) .setMaxSubwordSize(15) .setName(generateSimpleName(i++)), - new EdgeNGramTokenFilterV2() + new EdgeNGramTokenFilter() .setMinGram(1) .setMaxGram(2) .setSide(EdgeNGramTokenFilterSide.FRONT) @@ -1259,7 +1262,7 @@ SearchIndex createExpectedIndexWithSpecialDefaults(SearchIndex index) { new LimitTokenFilter() .setMaxTokenCount(1) .setName(generateSimpleName(i++)), - new NGramTokenFilterV2() + new NGramTokenFilter() .setMinGram(1) .setMaxGram(2) .setName(generateSimpleName(i++)), From 0694094d9ef15cd2629a78a9704470c919518929 Mon Sep 17 00:00:00 2001 From: Sima Zhu Date: Fri, 29 May 2020 12:23:23 -0700 Subject: [PATCH 3/4] Revert internal model changes --- .../indexes/implementation/models/EdgeNGramTokenFilter.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java index c5453429eee0..5149d1215862 100644 --- a/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java +++ b/sdk/search/azure-search-documents/src/main/java/com/azure/search/documents/indexes/implementation/models/EdgeNGramTokenFilter.java @@ -16,7 +16,7 @@ * of an input token. This token filter is implemented using Apache Lucene. */ @JsonTypeInfo(use = JsonTypeInfo.Id.NAME, include = JsonTypeInfo.As.PROPERTY, property = "@odata.type") -@JsonTypeName("#Microsoft.Azure.Search.EdgeNGramTokenFilterV2") +@JsonTypeName("#Microsoft.Azure.Search.EdgeNGramTokenFilter") @Fluent public final class EdgeNGramTokenFilter extends TokenFilter { /* From f111d36f7592a385dcc8c82fa2cc13721660c9ec Mon Sep 17 00:00:00 2001 From: Sima Zhu Date: Fri, 29 May 2020 12:31:16 -0700 Subject: [PATCH 4/4] Made change log changes --- sdk/search/azure-search-documents/CHANGELOG.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/sdk/search/azure-search-documents/CHANGELOG.md b/sdk/search/azure-search-documents/CHANGELOG.md index 1bde75f7e015..43a311a2d749 100644 --- a/sdk/search/azure-search-documents/CHANGELOG.md +++ b/sdk/search/azure-search-documents/CHANGELOG.md @@ -14,6 +14,10 @@ and uplifted the properties to `SearchIndexerDataSourceConnection` and `SearchResourceEncryptionKey` respectively. - Removed `select` parameter from list service resource APIs. - Added list names APIs for each search service resource. (e.g. `listSearchIndexNames`, `listSearchIndexerNames`, `listDataSourceNames`, `listSkillsetNames`, `listSynonymMapNames`) +- Removed deprecated versions and removed the V2 suffix. SDK is currently having `EdgeNGramTokenFilter`, `KeywordTokenizer`, `LuceneStandardTokenizer`, +`NGramTokenFilter`, and `PathHierarchyTokenizer`. +- Renamed `Similarity` to `SimilarityAlgorithm`. +- Renamed `Suggester` to `SearchSuggester`. ## 1.0.0-beta.3 (2020-05-05) - Replaced `isRetrievable` API with `isHidden`, parameter name changed from `retrievable` to `hidden`.