Azure.Search (preview:2024-11-01)

2025/01/09 • 5 updated methods

Indexes_Create (updated)
Description Creates a new search index.
Reference Link ¶

⚶ Changes

{
  "#id": "Indexes_Create",
  "$parameters": {
    "index": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    }
  },
  "$responses": {
    "201": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    }
  }
}

⚼ Request

POST:  /indexes
{
index:
{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}
,
x-ms-client-request-id: string ,
api-version: string ,
}

⚐ Response (201)

{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}

⚐ Response (default)

{
error:
{
code: string ,
message: string ,
target: string ,
details:
[
string ,
]
,
additionalInfo:
[
{
type: string ,
info: object ,
}
,
]
,
}
,
}
Indexes_List (updated)
Description Lists all indexes available for a search service.
Reference Link ¶

⚶ Changes

{
  "#id": "Indexes_List",
  "$responses": {
    "200": {
      "$properties": {
        "value": {
          "$properties": {
            "vectorSearch": [
              {
                "vectorizers": {
                  "$properties": [
                    {
                      "#name": "kind",
                      "Enum": {
                        "new": [
                          [
                            "azureOpenAI",
                            "Generate embeddings using an Azure OpenAI resource at query time."
                          ],
                          [
                            "customWebApi",
                            "Generate embeddings using a custom web endpoint at query time."
                          ],
                          [
                            "aiServicesVision",
                            "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                          ],
                          [
                            "aml",
                            "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                          ]
                        ],
                        "old": [
                          [
                            "azureOpenAI",
                            "Generate embeddings using an Azure OpenAI resource at query time."
                          ],
                          [
                            "customWebApi",
                            "Generate embeddings using a custom web endpoint at query time."
                          ],
                          [
                            "aiServicesVision",
                            "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                          ],
                          [
                            "aml",
                            "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                          ]
                        ]
                      }
                    }
                  ]
                }
              }
            ]
          }
        }
      }
    }
  }
}

⚼ Request

GET:  /indexes
{
$select: string ,
x-ms-client-request-id: string ,
api-version: string ,
}

⚐ Response (200)

{
value:
[
{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}
,
]
,
}

⚐ Response (default)

{
error:
{
code: string ,
message: string ,
target: string ,
details:
[
string ,
]
,
additionalInfo:
[
{
type: string ,
info: object ,
}
,
]
,
}
,
}
Indexes_CreateOrUpdate (updated)
Description Creates a new search index or updates an index if it already exists.
Reference Link ¶

⚶ Changes

{
  "#id": "Indexes_CreateOrUpdate",
  "$parameters": {
    "index": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    }
  },
  "$responses": {
    "200": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    },
    "201": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    }
  }
}

⚼ Request

PUT:  /indexes('{indexName}')
{
indexName: string ,
index:
{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}
,
allowIndexDowntime: boolean ,
x-ms-client-request-id: string ,
If-Match: string ,
If-None-Match: string ,
Prefer: string ,
api-version: string ,
}

⚐ Response (200)

{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}

⚐ Response (201)

{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}

⚐ Response (default)

{
error:
{
code: string ,
message: string ,
target: string ,
details:
[
string ,
]
,
additionalInfo:
[
{
type: string ,
info: object ,
}
,
]
,
}
,
}
Indexes_Get (updated)
Description Retrieves an index definition.
Reference Link ¶

⚶ Changes

{
  "#id": "Indexes_Get",
  "$responses": {
    "200": {
      "$properties": {
        "vectorSearch": [
          {
            "vectorizers": {
              "$properties": [
                {
                  "#name": "kind",
                  "Enum": {
                    "new": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog at query time."
                      ]
                    ],
                    "old": [
                      [
                        "azureOpenAI",
                        "Generate embeddings using an Azure OpenAI resource at query time."
                      ],
                      [
                        "customWebApi",
                        "Generate embeddings using a custom web endpoint at query time."
                      ],
                      [
                        "aiServicesVision",
                        "Generate embeddings for an image or text input at query time using the Azure AI Services Vision Vectorize API."
                      ],
                      [
                        "aml",
                        "Generate embeddings using an Azure Machine Learning endpoint deployed via the Azure AI Studio Model Catalog at query time."
                      ]
                    ]
                  }
                }
              ]
            }
          }
        ]
      }
    }
  }
}

⚼ Request

GET:  /indexes('{indexName}')
{
indexName: string ,
x-ms-client-request-id: string ,
api-version: string ,
}

⚐ Response (200)

{
name: string ,
fields:
[
{
name: string ,
type: enum ,
key: boolean ,
retrievable: boolean ,
stored: boolean ,
searchable: boolean ,
filterable: boolean ,
sortable: boolean ,
facetable: boolean ,
analyzer: enum ,
searchAnalyzer: enum ,
indexAnalyzer: enum ,
normalizer: enum ,
dimensions: integer ,
vectorSearchProfile: string ,
vectorEncoding: enum ,
synonymMaps:
[
string ,
]
,
fields:
[
string ,
]
,
}
,
]
,
scoringProfiles:
[
{
name: string ,
text:
{
weights: object ,
}
,
functions:
[
{
type: string ,
fieldName: string ,
boost: number ,
interpolation: enum ,
}
,
]
,
functionAggregation: enum ,
}
,
]
,
defaultScoringProfile: string ,
corsOptions:
{
allowedOrigins:
[
string ,
]
,
maxAgeInSeconds: integer ,
}
,
suggesters:
[
{
name: string ,
searchMode: enum ,
sourceFields:
[
string ,
]
,
}
,
]
,
analyzers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
tokenFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
charFilters:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
normalizers:
[
{
@odata.type: string ,
name: string ,
}
,
]
,
encryptionKey:
{
keyVaultKeyName: string ,
keyVaultKeyVersion: string ,
keyVaultUri: string ,
accessCredentials:
{
applicationId: string ,
applicationSecret: string ,
}
,
identity:
{
@odata.type: string ,
}
,
}
,
similarity:
{
@odata.type: string ,
}
,
semantic:
{
defaultConfiguration: string ,
configurations:
[
{
name: string ,
prioritizedFields:
{
titleField:
{
fieldName: string ,
}
,
prioritizedContentFields:
[
{
fieldName: string ,
}
,
]
,
prioritizedKeywordsFields:
[
{
fieldName: string ,
}
,
]
,
}
,
}
,
]
,
}
,
vectorSearch:
{
profiles:
[
{
name: string ,
algorithm: string ,
vectorizer: string ,
compression: string ,
}
,
]
,
algorithms:
[
{
name: string ,
kind: enum ,
}
,
]
,
vectorizers:
[
{
name: string ,
kind: enum ,
}
,
]
,
compressions:
[
{
name: string ,
kind: enum ,
rerankWithOriginalVectors: boolean ,
defaultOversampling: number ,
rescoringOptions:
{
enableRescoring: boolean ,
defaultOversampling: number ,
rescoreStorageMethod: enum ,
}
,
truncationDimension: integer ,
}
,
]
,
}
,
@odata.etag: string ,
}

⚐ Response (default)

{
error:
{
code: string ,
message: string ,
target: string ,
details:
[
string ,
]
,
additionalInfo:
[
{
type: string ,
info: object ,
}
,
]
,
}
,
}
Indexes_Analyze (updated)
Description Shows how an analyzer breaks text into tokens.
Reference Link ¶

⚶ Changes

{
  "#id": "Indexes_Analyze",
  "$parameters": {
    "request": {
      "$properties": {
        "tokenFilters": {
          "Enum": {
            "new": [
              [
                "arabic_normalization",
                "A token filter that applies the Arabic normalizer to normalize the orthography. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.html"
              ],
              [
                "apostrophe",
                "Strips all characters after an apostrophe (including the apostrophe itself). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/tr/ApostropheFilter.html"
              ],
              [
                "asciifolding",
                "Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the \"Basic Latin\" Unicode block) into their ASCII equivalents, if such equivalents exist. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ASCIIFoldingFilter.html"
              ],
              [
                "cjk_bigram",
                "Forms bigrams of CJK terms that are generated from the standard tokenizer. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKBigramFilter.html"
              ],
              [
                "cjk_width",
                "Normalizes CJK width differences. Folds full-width ASCII variants into the equivalent basic Latin, and half-width Katakana variants into the equivalent Kana. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKWidthFilter.html"
              ],
              [
                "classic",
                "Removes English possessives, and dots from acronyms. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/standard/ClassicFilter.html"
              ],
              [
                "common_grams",
                "Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/commongrams/CommonGramsFilter.html"
              ],
              [
                "edgeNGram_v2",
                "Generates n-grams of the given size(s) starting from the front or the back of an input token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.html"
              ],
              [
                "elision",
                "Removes elisions. For example, \"l'avion\" (the plane) will be converted to \"avion\" (plane). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/util/ElisionFilter.html"
              ],
              [
                "german_normalization",
                "Normalizes German characters according to the heuristics of the German2 snowball algorithm. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/de/GermanNormalizationFilter.html"
              ],
              [
                "hindi_normalization",
                "Normalizes text in Hindi to remove some differences in spelling variations. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/hi/HindiNormalizationFilter.html"
              ],
              [
                "indic_normalization",
                "Normalizes the Unicode representation of text in Indian languages. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/in/IndicNormalizationFilter.html"
              ],
              [
                "keyword_repeat",
                "Emits each incoming token twice, once as keyword and once as non-keyword. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/KeywordRepeatFilter.html"
              ],
              [
                "kstem",
                "A high-performance kstem filter for English. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/en/KStemFilter.html"
              ],
              [
                "length",
                "Removes words that are too long or too short. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/LengthFilter.html"
              ],
              [
                "limit",
                "Limits the number of tokens while indexing. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/LimitTokenCountFilter.html"
              ],
              [
                "lowercase",
                "Normalizes token text to lower case. See https://lucene.apache.org/core/6_6_1/analyzers-common/org/apache/lucene/analysis/core/LowerCaseFilter.html"
              ],
              [
                "nGram_v2",
                "Generates n-grams of the given size(s). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/NGramTokenFilter.html"
              ],
              [
                "persian_normalization",
                "Applies normalization for Persian. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/fa/PersianNormalizationFilter.html"
              ],
              [
                "phonetic",
                "Create tokens for phonetic matches. See https://lucene.apache.org/core/4_10_3/analyzers-phonetic/org/apache/lucene/analysis/phonetic/package-tree.html"
              ],
              [
                "porter_stem",
                "Uses the Porter stemming algorithm to transform the token stream. See http://tartarus.org/~martin/PorterStemmer"
              ],
              [
                "reverse",
                "Reverses the token string. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/reverse/ReverseStringFilter.html"
              ],
              [
                "scandinavian_normalization",
                "Normalizes use of the interchangeable Scandinavian characters. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ScandinavianNormalizationFilter.html"
              ],
              [
                "scandinavian_folding",
                "Folds Scandinavian characters \u00e5\u00c5\u00e4\u00e6\u00c4\u00c6->a and \u00f6\u00d6\u00f8\u00d8->o. It also discriminates against use of double vowels aa, ae, ao, oe and oo, leaving just the first one. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ScandinavianFoldingFilter.html"
              ],
              [
                "shingle",
                "Creates combinations of tokens as a single token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/shingle/ShingleFilter.html"
              ],
              [
                "snowball",
                "A filter that stems words using a Snowball-generated stemmer. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/snowball/SnowballFilter.html"
              ],
              [
                "sorani_normalization",
                "Normalizes the Unicode representation of Sorani text. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ckb/SoraniNormalizationFilter.html"
              ],
              [
                "stemmer",
                "Language specific stemming filter. See https://learn.microsoft.com/rest/api/searchservice/Custom-analyzers-in-Azure-Search#TokenFilters"
              ],
              [
                "stopwords",
                "Removes stop words from a token stream. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/core/StopFilter.html"
              ],
              [
                "trim",
                "Trims leading and trailing whitespace from tokens. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/TrimFilter.html"
              ],
              [
                "truncate",
                "Truncates the terms to a specific length. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/TruncateTokenFilter.html"
              ],
              [
                "unique",
                "Filters out tokens with same text as the previous token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/RemoveDuplicatesTokenFilter.html"
              ],
              [
                "uppercase",
                "Normalizes token text to upper case. See https://lucene.apache.org/core/6_6_1/analyzers-common/org/apache/lucene/analysis/core/UpperCaseFilter.html"
              ],
              [
                "word_delimiter",
                "Splits words into subwords and performs optional transformations on subword groups."
              ]
            ],
            "old": [
              [
                "arabic_normalization",
                "A token filter that applies the Arabic normalizer to normalize the orthography. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.html"
              ],
              [
                "apostrophe",
                "Strips all characters after an apostrophe (including the apostrophe itself). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/tr/ApostropheFilter.html"
              ],
              [
                "asciifolding",
                "Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the \"Basic Latin\" Unicode block) into their ASCII equivalents, if such equivalents exist. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ASCIIFoldingFilter.html"
              ],
              [
                "cjk_bigram",
                "Forms bigrams of CJK terms that are generated from the standard tokenizer. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKBigramFilter.html"
              ],
              [
                "cjk_width",
                "Normalizes CJK width differences. Folds fullwidth ASCII variants into the equivalent basic Latin, and half-width Katakana variants into the equivalent Kana. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/cjk/CJKWidthFilter.html"
              ],
              [
                "classic",
                "Removes English possessives, and dots from acronyms. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/standard/ClassicFilter.html"
              ],
              [
                "common_grams",
                "Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/commongrams/CommonGramsFilter.html"
              ],
              [
                "edgeNGram_v2",
                "Generates n-grams of the given size(s) starting from the front or the back of an input token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.html"
              ],
              [
                "elision",
                "Removes elisions. For example, \"l'avion\" (the plane) will be converted to \"avion\" (plane). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/util/ElisionFilter.html"
              ],
              [
                "german_normalization",
                "Normalizes German characters according to the heuristics of the German2 snowball algorithm. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/de/GermanNormalizationFilter.html"
              ],
              [
                "hindi_normalization",
                "Normalizes text in Hindi to remove some differences in spelling variations. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/hi/HindiNormalizationFilter.html"
              ],
              [
                "indic_normalization",
                "Normalizes the Unicode representation of text in Indian languages. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/in/IndicNormalizationFilter.html"
              ],
              [
                "keyword_repeat",
                "Emits each incoming token twice, once as keyword and once as non-keyword. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/KeywordRepeatFilter.html"
              ],
              [
                "kstem",
                "A high-performance kstem filter for English. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/en/KStemFilter.html"
              ],
              [
                "length",
                "Removes words that are too long or too short. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/LengthFilter.html"
              ],
              [
                "limit",
                "Limits the number of tokens while indexing. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/LimitTokenCountFilter.html"
              ],
              [
                "lowercase",
                "Normalizes token text to lower case. See https://lucene.apache.org/core/6_6_1/analyzers-common/org/apache/lucene/analysis/core/LowerCaseFilter.html"
              ],
              [
                "nGram_v2",
                "Generates n-grams of the given size(s). See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ngram/NGramTokenFilter.html"
              ],
              [
                "persian_normalization",
                "Applies normalization for Persian. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/fa/PersianNormalizationFilter.html"
              ],
              [
                "phonetic",
                "Create tokens for phonetic matches. See https://lucene.apache.org/core/4_10_3/analyzers-phonetic/org/apache/lucene/analysis/phonetic/package-tree.html"
              ],
              [
                "porter_stem",
                "Uses the Porter stemming algorithm to transform the token stream. See http://tartarus.org/~martin/PorterStemmer"
              ],
              [
                "reverse",
                "Reverses the token string. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/reverse/ReverseStringFilter.html"
              ],
              [
                "scandinavian_normalization",
                "Normalizes use of the interchangeable Scandinavian characters. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ScandinavianNormalizationFilter.html"
              ],
              [
                "scandinavian_folding",
                "Folds Scandinavian characters \u00e5\u00c5\u00e4\u00e6\u00c4\u00c6->a and \u00f6\u00d6\u00f8\u00d8->o. It also discriminates against use of double vowels aa, ae, ao, oe and oo, leaving just the first one. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/ScandinavianFoldingFilter.html"
              ],
              [
                "shingle",
                "Creates combinations of tokens as a single token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/shingle/ShingleFilter.html"
              ],
              [
                "snowball",
                "A filter that stems words using a Snowball-generated stemmer. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/snowball/SnowballFilter.html"
              ],
              [
                "sorani_normalization",
                "Normalizes the Unicode representation of Sorani text. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/ckb/SoraniNormalizationFilter.html"
              ],
              [
                "stemmer",
                "Language specific stemming filter. See https://learn.microsoft.com/rest/api/searchservice/Custom-analyzers-in-Azure-Search#TokenFilters"
              ],
              [
                "stopwords",
                "Removes stop words from a token stream. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/core/StopFilter.html"
              ],
              [
                "trim",
                "Trims leading and trailing whitespace from tokens. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/TrimFilter.html"
              ],
              [
                "truncate",
                "Truncates the terms to a specific length. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/TruncateTokenFilter.html"
              ],
              [
                "unique",
                "Filters out tokens with same text as the previous token. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/RemoveDuplicatesTokenFilter.html"
              ],
              [
                "uppercase",
                "Normalizes token text to upper case. See https://lucene.apache.org/core/6_6_1/analyzers-common/org/apache/lucene/analysis/core/UpperCaseFilter.html"
              ],
              [
                "word_delimiter",
                "Splits words into subwords and performs optional transformations on subword groups."
              ]
            ]
          }
        }
      }
    }
  }
}

⚼ Request

POST:  /indexes('{indexName}')/search.analyze
{
indexName: string ,
request:
{
text: string ,
analyzer: enum ,
tokenizer: enum ,
normalizer: enum ,
tokenFilters:
[
string ,
]
,
charFilters:
[
string ,
]
,
}
,
x-ms-client-request-id: string ,
api-version: string ,
}

⚐ Response (200)

{
tokens:
[
{
token: string ,
startOffset: integer ,
endOffset: integer ,
position: integer ,
}
,
]
,
}

⚐ Response (default)

{
error:
{
code: string ,
message: string ,
target: string ,
details:
[
string ,
]
,
additionalInfo:
[
{
type: string ,
info: object ,
}
,
]
,
}
,
}