You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/ai/conceptual/understanding-tokens.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -81,7 +81,7 @@ As training continues, the model adds any new tokens in the training text to its
81
81
82
82
The semantic relationships between the tokens can be analyzed by using these token ID sequences. Multi-valued numeric vectors, known as [embeddings](embeddings.md), are used to represent these relationships. An embedding is assigned to each token based on how commonly it's used together with, or in similar contexts to, the other tokens.
83
83
84
-
After it's trained, a model can calculate an embedding for text that contains multiple tokens. The model tokenizes the text, then calculates an overall embeddings value based on the learned embeddings of the individual tokens. This technique can be used for semantic document searches or adding [memories](/semantic-kernel/memories/) to an AI.
84
+
After it's trained, a model can calculate an embedding for text that contains multiple tokens. The model tokenizes the text, then calculates an overall embeddings value based on the learned embeddings of the individual tokens. This technique can be used for semantic document searches or adding [vector stores](/semantic-kernel/concepts/vector-store-connectors/) to an AI.
85
85
86
86
During output generation, the model predicts a vector value for the next token in the sequence. The model then selects the next token from it's vocabulary based on this vector value. In practice, the model calculates multiple vectors by using various elements of the previous tokens' embeddings. The model then evaluates all potential tokens from these vectors and selects the most probable one to continue the sequence.
This article demonstrates how to manage the connection between your App Service .NET application and a [vector database solution](../conceptual/vector-databases.md). It covers using Microsoft Entra managed identities for supported services and securely storing connection strings for others.
15
15
16
-
By adding a vector database to your application, you can enable [semantic memories](/semantic-kernel/memories/) for your AI. The [Semantic Kernel SDK](/semantic-kernel/overview) for .NET enables you to easily implement memory storage and recall using your preferred vector database solution.
16
+
By adding a vector database to your application, you can enable [semantic memories or *vector stores*]([vector stores](/semantic-kernel/concepts/vector-store-connectors/)) for your AI. The [Semantic Kernel SDK](/semantic-kernel/overview) for .NET enables you to easily implement memory storage and recall using your preferred vector database solution.
17
17
18
18
## Prerequisites
19
19
@@ -22,7 +22,7 @@ By adding a vector database to your application, you can enable [semantic memori
*[Create and deploy a .NET application to App Service](/azure/app-service/quickstart-dotnetcore)
25
-
*[Create and deploy a vector database solution](/semantic-kernel/memories/vector-db)
25
+
*[Create and deploy a vector database solution](/semantic-kernel/concepts/ai-services/integrations#vector-database-solutions)
26
26
27
27
## Use Microsoft Entra managed identity for authentication
28
28
@@ -190,7 +190,7 @@ Before following these steps, retrieve a connection string for your vector datab
190
190
> [!IMPORTANT]
191
191
> Before following these steps, ensure you have [created a Key Vault using the Azure CLI](/azure/key-vault/general/quick-create-cli).
192
192
193
-
1. Grant your user account permissions to your key vault through Role-Based Access Control (RBAC), assign a role using the Azure CLI command [`az role assignment create`](/cli/azure/role/assignment?view=azure-cli-latest#az-role-assignment-create):
193
+
1. Grant your user account permissions to your key vault through Role-Based Access Control (RBAC), assign a role using the Azure CLI command [`az role assignment create`](/cli/azure/role/assignment#az-role-assignment-create):
194
194
195
195
```azurecli
196
196
az role assignment create \
@@ -199,7 +199,7 @@ Before following these steps, retrieve a connection string for your vector datab
1. Add the connection string to Key Vault using the Azure CLI command [`az keyvault secret set`](/cli/azure/keyvault/secret?view=azure-cli-latest#az-keyvault-secret-set):
202
+
1. Add the connection string to Key Vault using the Azure CLI command [`az keyvault secret set`](/cli/azure/keyvault/secret#az-keyvault-secret-set):
203
203
204
204
```azurecli
205
205
az keyvault secret set \
@@ -271,7 +271,7 @@ Before following these steps, retrieve a connection string for your vector datab
271
271
272
272
:::zone target="docs" pivot="azure-cli"
273
273
274
-
Add or edit an app setting with the Azure CLI command [`az webapp config connection-string set`](/cli/azure/webapp/config/connection-string?view=azure-cli-latest#az-webapp-config-connection-string-set):
274
+
Add or edit an app setting with the Azure CLI command [`az webapp config connection-string set`](/cli/azure/webapp/config/connection-string#az-webapp-config-connection-string-set):
Copy file name to clipboardExpand all lines: docs/ai/how-to/use-redis-for-memory.md
+1-1
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@ ms.date: 04/17/2024
13
13
14
14
This article demonstrates how to integrate a Redis database with the RediSearch module into the [Semantic Kernel SDK](/semantic-kernel/overview) and use it for memory storage and retrieval.
15
15
16
-
[Memories](/semantic-kernel/memories/) represent text information that has been stored alongside a precomputed embedding vector for the whole text. When an LLM is prompted to recall a memory, it uses these precomputed embeddings to efficiently evaluate whether a memory is relevant to the prompt. After the LLM finds a matching memory, it uses the memory's text information as context for the next steps in the prompt completion.
16
+
[Vector stores](/semantic-kernel/concepts/vector-store-connectors/) represent text information that has been stored alongside a precomputed embedding vector for the whole text. When an LLM is prompted to recall a memory, it uses these precomputed embeddings to efficiently evaluate whether a memory is relevant to the prompt. After the LLM finds a matching memory, it uses the memory's text information as context for the next steps in the prompt completion.
17
17
18
18
Memory storage that's added to the Semantic Kernel SDK provides a broader context for your requests. It also enables you to store data in the same manner as you store a traditional database, but query it by using natural language.
Copy file name to clipboardExpand all lines: docs/ai/semantic-kernel-dotnet-overview.md
+10-10
Original file line number
Diff line number
Diff line change
@@ -15,7 +15,7 @@ In this article, you explore [Semantic Kernel](/semantic-kernel/overview) core c
15
15
- How to add semantic kernel to your project
16
16
- Semantic Kernel core concepts
17
17
18
-
The sections ahead serve as an introductory overview of Semantic Kernel specifically in the context of .NET. For more comprehensive information and training about Semantic Kernel, see the following resources:
18
+
This article serves as an introductory overview of Semantic Kernel specifically in the context of .NET. For more comprehensive information and training about Semantic Kernel, see the following resources:
@@ -178,21 +178,21 @@ The preceding code creates an executable, sequential plan to read content from a
178
178
179
179
### Memory
180
180
181
-
Semantic Kernel's [Memory](/semantic-kernel/memories) provides abstractions over embedding models, vector databases, and other data to simplify context management for AI applications. Memory is agnostic to the underlying LLM or Vector DB, offering a uniform developer experience. You can configure memory features to store data in a variety of sources or service, including Azure AI Search, Azure Cache for Redis, and more.
181
+
Semantic Kernel's [Vector stores](/semantic-kernel/concepts/vector-store-connectors/) provide abstractions over embedding models, vector databases, and other data to simplify context management for AI applications. Vector stores are agnostic to the underlying LLM or Vector database, offering a uniform developer experience. You can configure memory features to store data in a variety of sources or service, including Azure AI Search and Azure Cache for Redis.
0 commit comments