Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ public static com.azure.search.documents.indexes.implementation.models.CjkBigram
String name = obj.getName();
cjkBigramTokenFilter.setName(name);

Boolean outputUnigrams = obj.isOutputUnigrams();
Boolean outputUnigrams = obj.areOutputUnigrams();
cjkBigramTokenFilter.setOutputUnigrams(outputUnigrams);

if (obj.getIgnoreScripts() != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,10 +27,10 @@ public static CommonGramTokenFilter map(com.azure.search.documents.indexes.imple
commonGramTokenFilter.setName(name);

Boolean ignoreCase = obj.isIgnoreCase();
commonGramTokenFilter.setIgnoreCase(ignoreCase);
commonGramTokenFilter.setCaseIgnored(ignoreCase);

Boolean useQueryMode = obj.isUseQueryMode();
commonGramTokenFilter.setUseQueryMode(useQueryMode);
commonGramTokenFilter.setQueryModeUsed(useQueryMode);

if (obj.getCommonWords() != null) {
List<String> commonWords = new ArrayList<>(obj.getCommonWords());
Expand All @@ -53,10 +53,10 @@ public static com.azure.search.documents.indexes.implementation.models.CommonGra
String name = obj.getName();
commonGramTokenFilter.setName(name);

Boolean ignoreCase = obj.isIgnoreCase();
Boolean ignoreCase = obj.isCaseIgnored();
commonGramTokenFilter.setIgnoreCase(ignoreCase);

Boolean useQueryMode = obj.isUseQueryMode();
Boolean useQueryMode = obj.isQueryModeUsed();
commonGramTokenFilter.setUseQueryMode(useQueryMode);

if (obj.getCommonWords() != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ public static DictionaryDecompounderTokenFilter map(com.azure.search.documents.i
dictionaryDecompounderTokenFilter.setMinSubwordSize(minSubwordSize);

Boolean onlyLongestMatch = obj.isOnlyLongestMatch();
dictionaryDecompounderTokenFilter.setOnlyLongestMatch(onlyLongestMatch);
dictionaryDecompounderTokenFilter.setOnlyLongestMatched(onlyLongestMatch);

Integer maxSubwordSize = obj.getMaxSubwordSize();
dictionaryDecompounderTokenFilter.setMaxSubwordSize(maxSubwordSize);
Expand Down Expand Up @@ -61,7 +61,7 @@ public static com.azure.search.documents.indexes.implementation.models.Dictionar
Integer minSubwordSize = obj.getMinSubwordSize();
dictionaryDecompounderTokenFilter.setMinSubwordSize(minSubwordSize);

Boolean onlyLongestMatch = obj.isOnlyLongestMatch();
Boolean onlyLongestMatch = obj.isOnlyLongestMatched();
dictionaryDecompounderTokenFilter.setOnlyLongestMatch(onlyLongestMatch);

Integer maxSubwordSize = obj.getMaxSubwordSize();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ public static EntityRecognitionSkill map(com.azure.search.documents.indexes.impl
entityRecognitionSkill.setDescription(description);

Boolean includeTypelessEntities = obj.isIncludeTypelessEntities();
entityRecognitionSkill.setIncludeTypelessEntities(includeTypelessEntities);
entityRecognitionSkill.setTypelessEntitiesIncluded(includeTypelessEntities);

if (obj.getDefaultLanguageCode() != null) {
EntityRecognitionSkillLanguage defaultLanguageCode =
Expand Down Expand Up @@ -100,7 +100,7 @@ public static com.azure.search.documents.indexes.implementation.models.EntityRec
String description = obj.getDescription();
entityRecognitionSkill.setDescription(description);

Boolean includeTypelessEntities = obj.isIncludeTypelessEntities();
Boolean includeTypelessEntities = obj.areTypelessEntitiesIncluded();
entityRecognitionSkill.setIncludeTypelessEntities(includeTypelessEntities);

if (obj.getDefaultLanguageCode() != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,7 @@ public static com.azure.search.documents.indexes.implementation.models.KeepToken
keepTokenFilter.setKeepWords(keepWords);
}

Boolean lowerCaseKeepWords = obj.isLowerCaseKeepWords();
Boolean lowerCaseKeepWords = obj.areLowerCaseKeepWords();
keepTokenFilter.setLowerCaseKeepWords(lowerCaseKeepWords);
return keepTokenFilter;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ public static KeywordMarkerTokenFilter map(com.azure.search.documents.indexes.im
}

Boolean ignoreCase = obj.isIgnoreCase();
keywordMarkerTokenFilter.setIgnoreCase(ignoreCase);
keywordMarkerTokenFilter.setCaseIgnored(ignoreCase);
return keywordMarkerTokenFilter;
}

Expand All @@ -55,7 +55,7 @@ public static com.azure.search.documents.indexes.implementation.models.KeywordMa
keywordMarkerTokenFilter.setKeywords(keywords);
}

Boolean ignoreCase = obj.isIgnoreCase();
Boolean ignoreCase = obj.isCaseIgnored();
keywordMarkerTokenFilter.setIgnoreCase(ignoreCase);
return keywordMarkerTokenFilter;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ public static LimitTokenFilter map(com.azure.search.documents.indexes.implementa
limitTokenFilter.setMaxTokenCount(maxTokenCount);

Boolean consumeAllTokens = obj.isConsumeAllTokens();
limitTokenFilter.setConsumeAllTokens(consumeAllTokens);
limitTokenFilter.setAllTokensConsumed(consumeAllTokens);
return limitTokenFilter;
}

Expand All @@ -46,7 +46,7 @@ public static com.azure.search.documents.indexes.implementation.models.LimitToke
Integer maxTokenCount = obj.getMaxTokenCount();
limitTokenFilter.setMaxTokenCount(maxTokenCount);

Boolean consumeAllTokens = obj.isConsumeAllTokens();
Boolean consumeAllTokens = obj.areAllTokensConsumed();
limitTokenFilter.setConsumeAllTokens(consumeAllTokens);
return limitTokenFilter;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ public static MicrosoftLanguageStemmingTokenizer map(com.azure.search.documents.
}

Boolean isSearchTokenizer = obj.isSearchTokenizer();
microsoftLanguageStemmingTokenizer.setIsSearchTokenizer(isSearchTokenizer);
microsoftLanguageStemmingTokenizer.setIsSearchTokenizerUsed(isSearchTokenizer);
return microsoftLanguageStemmingTokenizer;
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ public static OcrSkill map(com.azure.search.documents.indexes.implementation.mod
}

Boolean shouldDetectOrientation = obj.isShouldDetectOrientation();
ocrSkill.shouldDetectOrientation(shouldDetectOrientation);
ocrSkill.setShouldDetectOrientation(shouldDetectOrientation);
return ocrSkill;
}

Expand Down Expand Up @@ -93,7 +93,7 @@ public static com.azure.search.documents.indexes.implementation.models.OcrSkill
ocrSkill.setDefaultLanguageCode(defaultLanguageCode);
}

Boolean shouldDetectOrientation = obj.shouldDetectOrientation();
Boolean shouldDetectOrientation = obj.setShouldDetectOrientation();
ocrSkill.setShouldDetectOrientation(shouldDetectOrientation);
return ocrSkill;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ public static PathHierarchyTokenizer map(com.azure.search.documents.indexes.impl
pathHierarchyTokenizer.setDelimiter(delimiter);

Boolean reverseTokenOrder = obj.isReverseTokenOrder();
pathHierarchyTokenizer.setReverseTokenOrder(reverseTokenOrder);
pathHierarchyTokenizer.setTokenOrderReversed(reverseTokenOrder);

Integer numberOfTokensToSkip = obj.getNumberOfTokensToSkip();
pathHierarchyTokenizer.setNumberOfTokensToSkip(numberOfTokensToSkip);
Expand Down Expand Up @@ -60,7 +60,7 @@ public static com.azure.search.documents.indexes.implementation.models.PathHiera
String delimiter = obj.getDelimiter();
pathHierarchyTokenizerV2.setDelimiter(delimiter);

Boolean reverseTokenOrder = obj.isReverseTokenOrder();
Boolean reverseTokenOrder = obj.isTokenOrderReversed();
pathHierarchyTokenizerV2.setReverseTokenOrder(reverseTokenOrder);

Integer numberOfTokensToSkip = obj.getNumberOfTokensToSkip();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ public static com.azure.search.documents.indexes.implementation.models.PatternAn
String name = obj.getName();
patternAnalyzer.setName(name);

Boolean lowerCaseTerms = obj.isLowerCaseTerms();
Boolean lowerCaseTerms = obj.areLowerCaseTerms();
patternAnalyzer.setLowerCaseTerms(lowerCaseTerms);

String pattern = obj.getPattern();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ public static PhoneticTokenFilter map(com.azure.search.documents.indexes.impleme
phoneticTokenFilter.setName(name);

Boolean replaceOriginalTokens = obj.isReplaceOriginalTokens();
phoneticTokenFilter.setReplaceOriginalTokens(replaceOriginalTokens);
phoneticTokenFilter.setOriginalTokensReplaced(replaceOriginalTokens);

if (obj.getEncoder() != null) {
PhoneticEncoder encoder = PhoneticEncoderConverter.map(obj.getEncoder());
Expand All @@ -48,7 +48,7 @@ public static com.azure.search.documents.indexes.implementation.models.PhoneticT
String name = obj.getName();
phoneticTokenFilter.setName(name);

Boolean replaceOriginalTokens = obj.isReplaceOriginalTokens();
Boolean replaceOriginalTokens = obj.areOriginalTokensReplaced();
phoneticTokenFilter.setReplaceOriginalTokens(replaceOriginalTokens);

if (obj.getEncoder() != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ public static RequestOptions map(com.azure.search.documents.implementation.model
RequestOptions requestOptions = new RequestOptions();

UUID xMsClientRequestId = obj.getXMsClientRequestId();
requestOptions.setXMsClientRequestId(xMsClientRequestId);
requestOptions.setClientRequestId(xMsClientRequestId);
return requestOptions;
}

Expand All @@ -36,7 +36,7 @@ public static com.azure.search.documents.implementation.models.RequestOptions ma
com.azure.search.documents.implementation.models.RequestOptions requestOptions =
new com.azure.search.documents.implementation.models.RequestOptions();

UUID xMsClientRequestId = obj.getXMsClientRequestId();
UUID xMsClientRequestId = obj.getClientRequestId();
requestOptions.setXMsClientRequestId(xMsClientRequestId);
return requestOptions;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ public static RequestOptions map(com.azure.search.documents.indexes.implementati
RequestOptions requestOptions = new RequestOptions();

UUID xMsClientRequestId = obj.getXMsClientRequestId();
requestOptions.setXMsClientRequestId(xMsClientRequestId);
requestOptions.setClientRequestId(xMsClientRequestId);
return requestOptions;
}

Expand All @@ -32,7 +32,7 @@ public static com.azure.search.documents.indexes.implementation.models.RequestOp
com.azure.search.documents.indexes.implementation.models.RequestOptions requestOptions =
new com.azure.search.documents.indexes.implementation.models.RequestOptions();

UUID xMsClientRequestId = obj.getXMsClientRequestId();
UUID xMsClientRequestId = obj.getClientRequestId();
requestOptions.setXMsClientRequestId(xMsClientRequestId);
return requestOptions;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,13 +60,13 @@ public static com.azure.search.documents.indexes.implementation.models.ShingleTo
Integer minShingleSize = obj.getMinShingleSize();
shingleTokenFilter.setMinShingleSize(minShingleSize);

Boolean outputUnigrams = obj.isOutputUnigrams();
Boolean outputUnigrams = obj.areOutputUnigrams();
shingleTokenFilter.setOutputUnigrams(outputUnigrams);

String filterToken = obj.getFilterToken();
shingleTokenFilter.setFilterToken(filterToken);

Boolean outputUnigramsIfNoShingles = obj.isOutputUnigramsIfNoShingles();
Boolean outputUnigramsIfNoShingles = obj.areOutputUnigramsIfNoShingles();
shingleTokenFilter.setOutputUnigramsIfNoShingles(outputUnigramsIfNoShingles);

Integer maxShingleSize = obj.getMaxShingleSize();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,10 +28,10 @@ public static StopwordsTokenFilter map(com.azure.search.documents.indexes.implem
stopwordsTokenFilter.setName(name);

Boolean removeTrailingStopWords = obj.isRemoveTrailingStopWords();
stopwordsTokenFilter.setRemoveTrailingStopWords(removeTrailingStopWords);
stopwordsTokenFilter.setTrailingStopWordsRemoved(removeTrailingStopWords);

Boolean ignoreCase = obj.isIgnoreCase();
stopwordsTokenFilter.setIgnoreCase(ignoreCase);
stopwordsTokenFilter.setCaseIgnored(ignoreCase);

if (obj.getStopwords() != null) {
List<String> stopwords = new ArrayList<>(obj.getStopwords());
Expand Down Expand Up @@ -59,10 +59,10 @@ public static com.azure.search.documents.indexes.implementation.models.Stopwords
String name = obj.getName();
stopwordsTokenFilter.setName(name);

Boolean removeTrailingStopWords = obj.isRemoveTrailingStopWords();
Boolean removeTrailingStopWords = obj.areTrailingStopWordsRemoved();
stopwordsTokenFilter.setRemoveTrailingStopWords(removeTrailingStopWords);

Boolean ignoreCase = obj.isIgnoreCase();
Boolean ignoreCase = obj.isCaseIgnored();
stopwordsTokenFilter.setIgnoreCase(ignoreCase);

if (obj.getStopwords() != null) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ public static SynonymTokenFilter map(com.azure.search.documents.indexes.implemen
}

Boolean ignoreCase = obj.isIgnoreCase();
synonymTokenFilter.setIgnoreCase(ignoreCase);
synonymTokenFilter.setCaseIgnored(ignoreCase);
return synonymTokenFilter;
}

Expand All @@ -53,15 +53,15 @@ public static com.azure.search.documents.indexes.implementation.models.SynonymTo
String name = obj.getName();
synonymTokenFilter.setName(name);

Boolean expand = obj.isExpand();
Boolean expand = obj.getExpand();
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should this be isExpand?

synonymTokenFilter.setExpand(expand);

if (obj.getSynonyms() != null) {
List<String> synonyms = new ArrayList<>(obj.getSynonyms());
synonymTokenFilter.setSynonyms(synonyms);
}

Boolean ignoreCase = obj.isIgnoreCase();
Boolean ignoreCase = obj.isCaseIgnored();
synonymTokenFilter.setIgnoreCase(ignoreCase);
return synonymTokenFilter;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ public static WordDelimiterTokenFilter map(com.azure.search.documents.indexes.im
wordDelimiterTokenFilter.setName(name);

Boolean catenateNumbers = obj.isCatenateNumbers();
wordDelimiterTokenFilter.setCatenateNumbers(catenateNumbers);
wordDelimiterTokenFilter.setNumbersCatenated(catenateNumbers);

if (obj.getProtectedWords() != null) {
List<String> protectedWords = new ArrayList<>(obj.getProtectedWords());
Expand Down Expand Up @@ -56,7 +56,7 @@ public static WordDelimiterTokenFilter map(com.azure.search.documents.indexes.im
wordDelimiterTokenFilter.setCatenateAll(catenateAll);

Boolean catenateWords = obj.isCatenateWords();
wordDelimiterTokenFilter.setCatenateWords(catenateWords);
wordDelimiterTokenFilter.setWordsCatenated(catenateWords);
return wordDelimiterTokenFilter;
}

Expand All @@ -74,7 +74,7 @@ public static com.azure.search.documents.indexes.implementation.models.WordDelim
String name = obj.getName();
wordDelimiterTokenFilter.setName(name);

Boolean catenateNumbers = obj.isCatenateNumbers();
Boolean catenateNumbers = obj.areNumbersCatenated();
wordDelimiterTokenFilter.setCatenateNumbers(catenateNumbers);

if (obj.getProtectedWords() != null) {
Expand All @@ -88,22 +88,22 @@ public static com.azure.search.documents.indexes.implementation.models.WordDelim
Boolean stemEnglishPossessive = obj.isStemEnglishPossessive();
wordDelimiterTokenFilter.setStemEnglishPossessive(stemEnglishPossessive);

Boolean splitOnCaseChange = obj.isSplitOnCaseChange();
Boolean splitOnCaseChange = obj.splitOnCaseChange();
wordDelimiterTokenFilter.setSplitOnCaseChange(splitOnCaseChange);

Boolean generateWordParts = obj.generateWordParts();
wordDelimiterTokenFilter.setGenerateWordParts(generateWordParts);

Boolean splitOnNumerics = obj.isSplitOnNumerics();
Boolean splitOnNumerics = obj.splitOnNumerics();
wordDelimiterTokenFilter.setSplitOnNumerics(splitOnNumerics);

Boolean preserveOriginal = obj.isPreserveOriginal();
wordDelimiterTokenFilter.setPreserveOriginal(preserveOriginal);

Boolean catenateAll = obj.isCatenateAll();
Boolean catenateAll = obj.catenateAll();
wordDelimiterTokenFilter.setCatenateAll(catenateAll);

Boolean catenateWords = obj.isCatenateWords();
Boolean catenateWords = obj.areWordsCatenated();
wordDelimiterTokenFilter.setCatenateWords(catenateWords);
return wordDelimiterTokenFilter;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@ public CjkBigramTokenFilter setIgnoreScripts(List<CjkBigramTokenFilterScripts> i
*
* @return the outputUnigrams value.
*/
public Boolean isOutputUnigrams() {
public Boolean areOutputUnigrams() {
return this.outputUnigrams;
}

Expand Down
Loading