Skip to content

Commit 54b8090

Browse files
authored
Merge pull request #6658 from EnterpriseDB/release-2025-03-31a
Release 2025-03-31a
2 parents d4ad3cc + b642a9f commit 54b8090

35 files changed

+375
-257
lines changed

advocacy_docs/edb-postgres-ai/ai-accelerator/capabilities.mdx

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -20,27 +20,27 @@ The storage locations can be used by AI Accelerator to create a volume. This vol
2020

2121
### Create a preparer (optional)
2222

23-
This step is optional and only needed for pre-processing source AI data.
23+
This step is optional and needed only for preprocessing source AI data.
2424

25-
A preparer is created with the Pipelines system. A preparer is a function that retrieves data from a table or volume and returns it in a format that can be used by the model.
25+
Create a preparer with the Pipelines system. A preparer is a function that retrieves data from a table or volume and returns it in a format that can be used by the model.
2626

27-
With any configuration, a Preparer requires:
27+
With any configuration, a preparer requires:
2828

29-
* a name
30-
* the data preparation operation
31-
* the name of the destination table where the output data will be stored
32-
* the name of the column in the destination table where the output data will be stored
29+
* A name
30+
* The data preparation operation
31+
* The name of the destination table where the output data will be stored
32+
* The name of the column in the destination table where the output data will be stored
3333

3434
If the preparer is for a table, it also needs:
3535

36-
* the name of the source table
37-
* the name of the column in the source table that contains the data
36+
* The name of the source table
37+
* The name of the column in the source table that contains the data
3838

39-
If, on the other hand, the preparer is for a volume, it needs:
39+
But if the preparer is for a volume, it needs:
4040

41-
* the name of the volume
41+
* The name of the volume
4242

43-
When a preparer is created, by default it assumes column identifiers of "id" for the source and destination but this can be customized.
43+
When a preparer is created, by default it assumes column identifiers of "id" for the source and destination, but this can be customized.
4444

4545
### Create a model
4646

advocacy_docs/edb-postgres-ai/ai-accelerator/gettingstarted/index.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -59,12 +59,12 @@ INSERT 0 9
5959
The result is a table with some data in it: food products and some very personal opinions about them.
6060

6161
!!! Note
62-
This example uses simple, processed data. Pipelines can also handle multiple pre-processing steps via [Preparers](../preparers).
62+
This example uses simple, processed data. Pipelines can also handle multiple preprocessing steps via [preparers](../preparers).
6363
!!!
6464

6565
## Creating a retriever
6666

67-
The core of Pipelines is the Retriever. A retriever is a way to access the data in the table and use it in AI workflows.
67+
The core of Pipelines is the *retriever*. A retriever is a way to access the data in the table and use it in AI workflows.
6868

6969
```sql
7070
select aidb.create_retriever_for_table('products_retriever', 't5', 'products', 'description', 'Text');
@@ -196,6 +196,6 @@ __OUTPUT__
196196

197197
In [Models](../models), you can learn how to create more models with Pipelines, including external models from OpenAI API-compatible services.
198198

199-
In [Preparers](../preparers), you can learn more about how to create Preparers in order to perform pre-processing steps on your data before embeddings are calculated.
199+
In [Preparers](../preparers), you can learn more about how to create preparers to perform preprocessing steps on your data before embeddings are calculated.
200200

201201
In [Retrievers](../retrievers), you can learn more about how to use retrievers with external data sources, local files, or S3 storage and how to use the retriever functions to get the data you need.

advocacy_docs/edb-postgres-ai/ai-accelerator/limitations.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -33,13 +33,13 @@ There's currently no load balancing mechanism for model access.
3333

3434
Retriever Pipelines currently supports only text and image formats. Other formats, including structured data, video, and audio, aren't currently supported.
3535

36-
### Auto-Embedding for Images
36+
### Auto-embedding for images
3737

38-
* Auto-Embedding is only supported for text data. Image data embeddings can be manually computed.
38+
* Auto-embedding is supported only for text data. Image data embeddings can be manually computed.
3939

40-
### Auto-Data Preparation
40+
### Auto-data preparation
4141

42-
* Auto-Data Preparation is not currently supported. Bulk data preparation can be executed manually.
42+
* Auto-data preparation isn't currently supported. Bulk data preparation can be executed manually.
4343

4444
## Upgrading
4545

advocacy_docs/edb-postgres-ai/ai-accelerator/models/index.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,6 @@ Pipelines has a model registry that manages configured instances of models. Any
1717
* See the [supported models](./supported-models) that come with Pipelines.
1818
* Using [models with OpenAI API-compatible services and Nvidia NIM](using-with) with Pipelines.
1919

20-
## Next Steps
20+
## Next steps
2121

2222
Once you are familiar with models, you can learn how to use those models with [retrievers](../retrievers).

advocacy_docs/edb-postgres-ai/ai-accelerator/models/supported-models/completions.mdx

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -13,9 +13,7 @@ Model aliases:
1313

1414
## About completions
1515

16-
Completions enables the use of any OpenAI API-compatible text-generation model.
17-
18-
It's suitable for chat/text transforms, text completion, and other text generation tasks.
16+
Completions enables the use of any OpenAI API-compatible text-generation model. It's suitable for chat/text transforms, text completion, and other text generation tasks.
1917

2018
Based on the name of the model, the model provider sets defaults accordingly:
2119

@@ -99,7 +97,7 @@ The following configuration settings are available for OpenAI models:
9997

10098
## Model credentials
10199

102-
The following credentials may be required by the service providing these models. Note: `api_key` and `basic_auth` are exclusive. Only one of these two options can be used.
100+
The following credentials may be required by the service providing these models. Note: `api_key` and `basic_auth` are exclusive. You can use only one of these two options.
103101

104-
* `api_key` &mdash; The API key to use for Bearer Token authentication. The api_key will be sent in a header field as `Authorization: Bearer <api_key>`.
105-
* `basic_auth` &mdash; credentials for HTTP Basic authentication. The credentials provided here will be sent verbatim as `Authorization: Basic <basic_auth>`. If your server requires username/password with HTTP Basic authentication, use the syntax `username:password`. If only a token is required, only provide the token.
102+
* `api_key` &mdash; The API key to use for Bearer Token authentication. The api_key is sent in a header field as `Authorization: Bearer <api_key>`.
103+
* `basic_auth` &mdash; Credentials for HTTP Basic authentication. The credentials provided here are sent verbatim as `Authorization: Basic <basic_auth>`. If your server requires username/password with HTTP Basic authentication, use the syntax `username:password`. If only a token is required, provide only the token.

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-models.mdx

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ Pipelines has a model registry that manages configured instances of models. Any
88

99
## Discover the preloaded models
1010

11-
Pipelines comes with a set of pre-configured models that you can use out of the box.
11+
Pipelines comes with a set of preconfigured models that you can use out of the box.
1212

1313
To find them, run the following query:
1414

@@ -27,13 +27,13 @@ This query returns a list of all the models that are currently created in the sy
2727
dummy | dummy | {"config={}"}
2828
```
2929

30-
The `bert`, `clip`, and `t5` models are all pre-configured and ready to use. The `dummy` model is a placeholder model that can be used for testing purposes.
30+
The `bert`, `clip`, and `t5` models are all preconfigured and ready to use. The `dummy` model is a placeholder model that can be used for testing purposes.
3131

3232
!!! note First use of local models
33-
The first time you use any of the local models, the model will be downloaded from [HuggingFace](https://huggingface.co/). The model is then run locally.
33+
The first time you use any of the local models, the model will be downloaded from [HuggingFace](https://huggingface.co/). The model is then run locally.
3434
Subsequent uses of the model will be faster, as the model will be cached locally.
3535

36-
If you use a proxy, ensure that it is [configured on the Postgres server](../installing/complete#proxy-settings).
36+
If you use a proxy, ensure that it's [configured on the Postgres server](../installing/complete#proxy-settings).
3737

3838
!!!
3939

@@ -101,7 +101,7 @@ SELECT aidb.create_model(
101101
);
102102
```
103103

104-
Replace the `api_key` value with your own OpenAI API key (Read more about credential handling under [supported models](./supported-models)). Then you can use the `my_openai_model` model in your Pipelines functions and, in this example, leverage the GPT-4o model from OpenAI.
104+
Replace the `api_key` value with your own OpenAI API key. (For more information about credential handling, see [supported models](./supported-models).) Then you can use the `my_openai_model` model in your Pipelines functions and, in this example, leverage the GPT-4o model from OpenAI.
105105

106106
Creating the OpenAI embeddings model is similar to creating the completions model:
107107

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-with/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,8 @@ navTitle: Using models with...
44
description: How to use OpenAI-compatible and Nvidia NIM models with AI Accelerator Pipelines.
55
---
66

7-
In this section we look at some particular ways to use models with AI Accelerator Pipelines.
8-
These techniques show how to use API compatible services, running locally or in the cloud, with Pipelines and how to make use of Nvidia NIM models.
7+
This section describes some particular ways to use models with AI Accelerator Pipelines.
8+
These techniques show how to use API-compatible services, running locally or in the cloud, with Pipelines and how to make use of Nvidia NIM models.
99

1010
* [OpenAI API-compatible services](openai-api-compatibility) with Pipelines.
1111
* [Nvidia NIM models](using-nvidia-nim) with Pipelines.

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-with/openai-api-compatibility.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ Specify the model provider as `embeddings`, which is the provider that defaults
3737

3838
### Configuration
3939

40-
The next parameter is the configuration. This is a JSON string. When expanded, it has two parameters: the model, and the url.
40+
The next parameter is the configuration. This is a JSON string. When expanded, it has two parameters: the model and the url.
4141

4242
```json
4343
'{"model":"llama3.2", "url":"http://llama.local:11434/v1/embeddings"}'::JSONB

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-with/using-nvidia-nim/index.mdx

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,9 @@ navTitle: Nvidia NIM models
44
description: How to use Nvidia NIM models, either in your own environment or in the Nvidia cloud, with AI Accelerator Pipelines.
55
---
66

7-
You can use Nvidia NIM models with AI Accelerator. The models can be run in the Nvidia cloud or run within your environment under your control.
7+
You can use Nvidia NIM models with AI Accelerator. The models can run in the Nvidia cloud or in your environment under your control.
88

9-
This guide shows you how to use Nvidia NIM models with AI Accelerator in both scenarios.
9+
You can learn how to use Nvidia NIM models with AI Accelerator in both scenarios:
1010

1111
* [In the Nvidia cloud](using-nim-in-nvidia-cloud)
1212
* [In your environment](using-nim-in-your-environment)

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-with/using-nvidia-nim/using-nim-in-nvidia-cloud.mdx

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -5,37 +5,37 @@ description: Learn how to use Nvidia NIM models in the Nvidia cloud, hosted by N
55
---
66

77

8-
To use a Nvidia NIM that is hosted Nvidia's cloud, you first need a select a model to use. For this tutorial, we will be using the Nvidia NIM model llama-3.3-70b-instruct.
8+
To use a Nvidia NIM that's hosted in Nvidia's cloud, you first need to select a model to use. This tutorial uses the Nvidia NIM model llama-3.3-70b-instruct.
99

1010
## Prerequisites
1111

12-
* An Nvidia NGC account. (If you don't have one, you can create one [here](https://build.nvidia.com/explore/discover/))
12+
* An Nvidia NGC account. (If you don't have one, you can create one [here](https://build.nvidia.com/explore/discover/).)
1313

1414
## Configuring the Nvidia cloud
1515

16-
### 1. Select a Model
16+
### 1. Select a model
1717

18-
Choose a model from [Nvidia's model library](https://build.nvidia.com/models). For this example, we will used the [llama-3.3-70b-instruct](https://build.nvidia.com/meta/llama3-70b) model.
18+
Choose a model from [Nvidia's model library](https://build.nvidia.com/models). This example uses the [llama-3.3-70b-instruct](https://build.nvidia.com/meta/llama3-70b) model.
1919

2020
### 2. Generate an API Key
2121

22-
API keys can be created from the model’s page.
22+
You can create API keys from the model’s page.
2323

2424
![Get API Key](images/get-api-key.svg)
2525

26-
Click on the `Get API Key` button to generate an API key. If you have not logged in, you will be prompted to log in and if you have not signed up, you will be invited to do so. You will recieve an API key that you can use to interact with the model. We will refer to this key as the `<NIM API KEY>` in the following steps.
26+
To generate an API key, select **Get API Key**. If you haven't logged in, you're prompted to log in. If you haven't signed up, you're invited to do so. You'll receive an API key that you can use to interact with the model. This example refers to this key as the `<NIM API KEY>`.
2727

28-
## Integrating the Model with AI Accelerator
28+
## Integrating the model with AI Accelerator
2929

3030
### 1. Enable AI Accelerator in EDB Postgres AI
3131

32-
With an EDB Postgres Advanced Server, EDB Postgres Extended Server, or a Community Postgres instance running, connect to the database and enable the AI Accelerator extension:
32+
With EDB Postgres Advanced Server, EDB Postgres Extended Server, or a community Postgres instance running, connect to the database and enable the AI Accelerator extension:
3333

3434
```sql
3535
CREATE EXTENSION aidb CASCADE;
3636
```
3737

38-
### 2. Register the Model
38+
### 2. Register the model
3939

4040
```sql
4141
SELECT aidb.create_model( \
@@ -46,9 +46,9 @@ SELECT aidb.create_model( \
4646
);
4747
```
4848

49-
### 3. Run the Model
49+
### 3. Run the model
5050

51-
Execute the following query to interact with the model:
51+
To interact with the model, execute the following query:
5252

5353
```sql
5454
SELECT aidb.decode_text('my_nim_llm', 'Tell me a short, one sentence story');
@@ -58,4 +58,4 @@ __OUTPUT__
5858
As the clock struck midnight, a single tear fell from the porcelain doll's glassy eye.
5959
```
6060
61-
Your output may vary. You have successfully used Nvidia NIM models, running on Nvidia's cloud, integrated with AI Accelerator.
61+
Your output may vary. You've successfully used Nvidia NIM models, running on Nvidia's cloud, integrated with AI Accelerator.

advocacy_docs/edb-postgres-ai/ai-accelerator/models/using-with/using-nvidia-nim/using-nim-in-your-environment.mdx

Lines changed: 20 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -4,12 +4,12 @@ navTitle: In your environment
44
description: Learn how to use Nvidia NIM models in your environment, locally or in your private cloud.
55
---
66

7-
To use a Nvidia NIM that is hosted in your own environment, you first need a instance of the model. For this tutorial, we will show how to configure an AWS hosted instance with the Nvidia NIM model. We will use the Nvidia NIM model llama3-8b-instruct.
7+
To use a Nvidia NIM that's hosted in your own environment, you first need an instance of the model. This tutorial shows how to configure an AWS-hosted instance with the Nvidia NIM model. It uses the Nvidia NIM model llama3-8b-instruct.
88

99
## Prerequisites
1010

11-
* A system capable of running Nvidia CUDA Toolkit. We suggest, for this tutorial, using an **EC2 g5.8xlarge instance** with **1024 GB of gp3 storage** running **Ubuntu 24.04 LTS**, although smaller instance sizes may also work.
12-
* An Nvidia NGC account. (If you don't have one, you can create one [here](https://build.nvidia.com/explore/discover/))
11+
* A system capable of running Nvidia CUDA Toolkit. For this tutorial, we recommend using an **EC2 g5.8xlarge instance** with **1024 GB of gp3 storage** running **Ubuntu 24.04 LTS**, although smaller instance sizes may also work.
12+
* A Nvidia NGC account. (If you don't have one, you can create one [here](https://build.nvidia.com/explore/discover/).)
1313

1414
## Configuring the system
1515

@@ -19,15 +19,16 @@ Download and install the CUDA Toolkit from [Nvidia's official page](https://deve
1919

2020
### 2. Install Docker
2121

22-
If your system does not have Docker installed, download and install docker.
23-
Instructions for Ubuntu are [here](https://docs.docker.com/engine/install/ubuntu/).
24-
Instructions for [other Linux platforms are available](https://docs.docker.com/engine/install/) too.
22+
If your system doesn't have Docker installed, download and install it:
23+
24+
- Instructions for Ubuntu are [here](https://docs.docker.com/engine/install/ubuntu/).
25+
- Instructions for other Linux platforms are [here](https://docs.docker.com/engine/install/).
2526

2627
### 3. Generate an NGC API Key
2728

28-
Obtain an API key from [Nvidia NGC](https://org.ngc.nvidia.com/setup/api-key). This key will be referred to as the NGC API KEY in the following steps.
29+
Obtain an API key from [Nvidia NGC](https://org.ngc.nvidia.com/setup/api-key). This example refers to this key as `<NGC API KEY>`.
2930

30-
Login to the NGC container registry:
31+
Log in to the NGC container registry:
3132

3233
```shell
3334
docker login nvcr.io
@@ -44,7 +45,7 @@ Download and install the Nvidia NGC CLI from [here](https://org.ngc.nvidia.com/s
4445

4546
### 5. Run the NIM Model
4647

47-
Save the following script as a shell script and execute it (more information can be found from [Nvidia's documentation](https://docs.nvidia.com/nim/large-language-models/latest/getting-started.html#serving-models-from-local-assets)). Remember to substitute the `<NGC API KEY>` with the API key you generated:
48+
Save the following script as a shell script and execute it. (For more information, see [Nvidia's documentation](https://docs.nvidia.com/nim/large-language-models/latest/getting-started.html#serving-models-from-local-assets).) Remember to substitute `<NGC API KEY>` with the API key you generated.
4849

4950
```shell
5051
# Choose a container name
@@ -68,9 +69,9 @@ docker run -it --rm --name=$CONTAINER_NAME \
6869
$IMG_NAME
6970
```
7071

71-
### 6. Test the Deployment
72+
### 6. Test the deployment
7273

73-
Run the following command to verify the model is working:
74+
To verify the model is working, run the following command:
7475

7576
```shell
7677
curl -X POST "http://0.0.0.0:8000/v1/chat/completions" \
@@ -87,21 +88,21 @@ curl -X POST "http://0.0.0.0:8000/v1/chat/completions" \
8788
}'
8889
```
8990

90-
Once you have verified the model is working, you can use it with EDB Postgres AI Accelerator.
91+
Once you've verified the model is working, you can use it with EDB Postgres AI Accelerator.
9192

92-
Make a note of the public IP address of the EC2 instance, as you will need it to connect to the model from EDB Postgres AI Accelerator. This will be referred to as `<NIM_HOST>` in the following steps.
93+
Record the public IP address of the EC2 instance, as you'll need it to connect to the model from EDB Postgres AI Accelerator. This example refers to this address as `<NIM_HOST>`.
9394

94-
## Integrating the Model with AI Accelerator
95+
## Integrating the model with AI Accelerator
9596

9697
### 1. Enable AI Accelerator in EDB Postgres AI
9798

98-
With an EDB Postgres Advanced Server, EDB Postgres Extended Server, or a Community Postgres instance running, connect to the database and enable the AI Accelerator extension by running the following SQL commands:
99+
With EDB Postgres Advanced Server, EDB Postgres Extended Server, or a community Postgres instance running, connect to the database. Then enable the AI Accelerator extension by running the following SQL commands:
99100

100101
```sql
101102
CREATE EXTENSION aidb CASCADE;
102103
```
103104

104-
### 2. Register the Model
105+
### 2. Register the model
105106

106107
```sql
107108
SELECT aidb.create_model(
@@ -111,9 +112,9 @@ SELECT aidb.create_model(
111112
);
112113
```
113114

114-
### 3. Run the Model
115+
### 3. Run the model
115116

116-
Execute the following query to interact with the model:
117+
To interact with the model, execute the following query:
117118

118119
```sql
119120
SELECT aidb.decode_text('my_nim_llm', 'Tell me a short, one sentence story');
@@ -123,4 +124,4 @@ __OUTPUT__
123124
As the clock struck midnight, a single tear fell from the porcelain doll's glassy eye.
124125
```
125126
126-
Your output may vary. You have successfully used Nvidia NIM models via EDB’s AI Accelerator.
127+
Your output may vary. You've successfully used Nvidia NIM models via the EDB AI Accelerator.

0 commit comments

Comments
 (0)