diff --git a/src/_data/catalog/destination_categories.yml b/src/_data/catalog/destination_categories.yml index ca461149bd..643b9105e5 100644 --- a/src/_data/catalog/destination_categories.yml +++ b/src/_data/catalog/destination_categories.yml @@ -1,5 +1,5 @@ # AUTOGENERATED FROM PUBLIC API. DO NOT EDIT -# destination categories last updated 2025-04-10 +# destination categories last updated 2025-04-16 items: - display_name: A/B Testing slug: a-b-testing diff --git a/src/_data/catalog/destinations.yml b/src/_data/catalog/destinations.yml index 59414630d4..27ea7a85e2 100644 --- a/src/_data/catalog/destinations.yml +++ b/src/_data/catalog/destinations.yml @@ -1,5 +1,5 @@ # AUTOGENERATED FROM PUBLIC API. DO NOT EDIT -# destination data last updated 2025-04-10 +# destination data last updated 2025-04-16 items: - id: 637e8d185e2dec264895ea89 display_name: 1Flow diff --git a/src/_data/catalog/destinations_private.yml b/src/_data/catalog/destinations_private.yml index 61d530d36b..321738d13b 100644 --- a/src/_data/catalog/destinations_private.yml +++ b/src/_data/catalog/destinations_private.yml @@ -1,5 +1,5 @@ # AUTOGENERATED FROM PUBLIC API. DO NOT EDIT -# destination data last updated 2025-04-10 +# destination data last updated 2025-04-16 items: - id: 54521fd925e721e32a72eee1 display_name: Pardot diff --git a/src/_data/catalog/source_categories.yml b/src/_data/catalog/source_categories.yml index af303e8e13..9f2655d02f 100644 --- a/src/_data/catalog/source_categories.yml +++ b/src/_data/catalog/source_categories.yml @@ -1,5 +1,5 @@ # AUTOGENERATED FROM PUBLIC API. DO NOT EDIT -# source categories last updated 2025-04-10 +# source categories last updated 2025-04-16 items: - display_name: A/B Testing slug: a-b-testing diff --git a/src/_data/catalog/sources.yml b/src/_data/catalog/sources.yml index b05ab5cfeb..d99cc6d0ff 100644 --- a/src/_data/catalog/sources.yml +++ b/src/_data/catalog/sources.yml @@ -1,5 +1,5 @@ # AUTOGENERATED FROM PUBLIC API. DO NOT EDIT -# sources last updated 2025-04-10 +# sources last updated 2025-04-16 items: - id: 8HWbgPTt3k display_name: .NET diff --git a/src/connections/destinations/catalog/actions-klaviyo/index.md b/src/connections/destinations/catalog/actions-klaviyo/index.md index 0521bb9f3f..763ae25b4b 100644 --- a/src/connections/destinations/catalog/actions-klaviyo/index.md +++ b/src/connections/destinations/catalog/actions-klaviyo/index.md @@ -107,10 +107,6 @@ Some customers experience 403 errors when sending audience data to Klaviyo throu To reduce the number of `403` errors that you encounter, enable [IP Allowlisting](/docs/connections/destinations/#ip-allowlisting) for your workspace. For more information the range of IP addresses Klaviyo uses for integration traffic, see Klaviyo's [How to allowlist Klaviyo integration traffic IP addresses](https://help.klaviyo.com/hc/en-us/articles/19143781289115){:target="_blank”} documentation. -#### Can I send Engage Audiences to a pre-created Klaviyo List? - -No. Engage audiences are designed to initiate the creation of new lists in Klaviyo when you use the "Add Profile to List - Engage" mapping. You cannot link Engage lists to existing Klaviyo lists and cannot edit the List ID for Engage audiences. - #### How can I unsuppress a profile when adding it to a list? When adding a user to a list, our action make use of the [Bulk Profile Import](https://developers.klaviyo.com/en/reference/spawn_bulk_profile_import_job){:target="_blank”} endpoint (when batching is enabled), and the [Add Profile To List](https://developers.klaviyo.com/en/reference/create_list_relationships){:target="_blank”} endpoint for non-batched requests. Both of which will not update a users suppression status if they were previously suppressed. diff --git a/src/connections/destinations/catalog/actions-tiktok-audiences/index.md b/src/connections/destinations/catalog/actions-tiktok-audiences/index.md index 00b9b4b75c..c8a092e50a 100644 --- a/src/connections/destinations/catalog/actions-tiktok-audiences/index.md +++ b/src/connections/destinations/catalog/actions-tiktok-audiences/index.md @@ -23,7 +23,9 @@ By using Segment's TikTok Audiences destination, you can increase traffic and dr ### Prerequisites -Before connecting to the TikTok Audiences destination, you must have a [TikTok Ads Manager](https://www.tiktok.com/business/en-US/solutions/ads-manager){:target="_blank"} account. +Before connecting to the TikTok Audiences destination, you must have a [TikTok Ads Manager](https://www.tiktok.com/business/en-US/solutions/ads-manager){:target="_blank"} account, with either Admin or Operator permissions to create and manage campaigns in TikTok. + +For more details on account and access level permissions, refer to [TikTok's documentation](https://ads.tiktok.com/help/article/how-to-assign-asset-level-permissions?lang=en){:target="_blank"}. ### TikTok Audience Segments diff --git a/src/connections/destinations/catalog/facebook-offline-conversions/index.md b/src/connections/destinations/catalog/facebook-offline-conversions/index.md index 388c05465f..092d34e722 100644 --- a/src/connections/destinations/catalog/facebook-offline-conversions/index.md +++ b/src/connections/destinations/catalog/facebook-offline-conversions/index.md @@ -6,6 +6,9 @@ id: 58ae54dc70a3e552b95415f6 --- [Facebook Offline Conversions](https://www.facebook.com/business/help/1782327938668950?utm_source=segmentio&utm_medium=docs&utm_campaign=partners){:target="_blank”} enables offline event tracking, so marketers can run campaigns, upload transaction data, and compare in-store transactions. +> info "Offline Conversions API deprecation" +> Meta will [discontinue the Offline Conversions API](https://developers.facebook.com/docs/graph-api/changelog/version17.0#offline-conversions-api){:target="_blank"} in May 2025. As a result, this destination will stop accepting data at that time and will no longer be available for use. To continue sending offline conversion events to Meta, migrate to the [Facebook Conversions API (Actions)](/docs/connections/destinations/catalog/actions-facebook-conversions-api/#purchase) destination, which supports offline event tracking. + > info "Customer Information Parameters Requirements" > As of Facebook Marketing API v13.0+, Facebook began enforcing new requirements for customer information parameters (match keys). To ensure your events don't throw an error, Segment recommends that you review [Facebook’s new requirements](https://developers.facebook.com/docs/graph-api/changelog/version13.0#conversions-api){:target="_blank"}. diff --git a/src/connections/destinations/catalog/facebook-pixel/index.md b/src/connections/destinations/catalog/facebook-pixel/index.md index 5fb11ea0f1..579c00566a 100644 --- a/src/connections/destinations/catalog/facebook-pixel/index.md +++ b/src/connections/destinations/catalog/facebook-pixel/index.md @@ -110,7 +110,7 @@ In addition, Segment sends the following event types as Standard events: - `Products Searched`, which Segment sends as `Search` - `Checkout Started`, which Segment sends as `InitiateCheckout` -Facebook requires a currency for `Purchase` events. If you leave it out a currency, Segment will set a default value of `USD`. +Facebook requires a currency for `Purchase` events. If you leave out a currency, Segment will set a default value of `USD`. You can set custom properties for the events listed above. Use the setting "Standard Events custom properties" to list all the properties you want to send. @@ -193,7 +193,7 @@ If you're using real estate, travel, or automotive [Dynamic Ads](https://www.fac For most implementations, Segment recommends leaving these mappings blank. By default, Segment sets `content_type` to "product". -The same mapping can be used to change the `content_id` from the default value (product_id or the sku) to anything specific for Meta Pixel. For more information about required Meta Pixel events, see Meta's [Required Meta Pixel events and parameters for Advantage+ catalog ads](https://www.facebook.com/business/help/606577526529702?id=1205376682832142){:target="_blank”} documentation. +The same mapping can be used to change the `content_ids` from the default value (product_id or the sku) to anything specific for Meta Pixel. For more information about required Meta Pixel events, see Meta's [Required Meta Pixel events and parameters for Advantage+ catalog ads](https://www.facebook.com/business/help/606577526529702?id=1205376682832142){:target="_blank”} documentation. ## Troubleshooting diff --git a/src/connections/functions/destination-functions.md b/src/connections/functions/destination-functions.md index ae56dfd98f..f109a68a4a 100644 --- a/src/connections/functions/destination-functions.md +++ b/src/connections/functions/destination-functions.md @@ -79,6 +79,12 @@ To change which event type the handler listens to, you can rename it to the name > info "" > Functions' runtime includes a `fetch()` polyfill using a `node-fetch` package. Check out the [node-fetch documentation](https://www.npmjs.com/package/node-fetch){:target="_blank"} for usage examples. +### Variable scoping + +When declaring settings variables, declare them in the function handler rather than globally in your function. This prevents you from leaking the settings values across other function instances. + +The handler for destination functions is event-specific. For example, you might have an `onTrack()`or `onIdentify()` function handler. + ### Errors and error handling {% include content/functions/errors-and-error-handling.md %} diff --git a/src/connections/functions/insert-functions.md b/src/connections/functions/insert-functions.md index e364ba6c94..f40678d9df 100644 --- a/src/connections/functions/insert-functions.md +++ b/src/connections/functions/insert-functions.md @@ -111,6 +111,12 @@ To ensure the Destination processes an event payload modified by the function, r > info "" > Functions' runtime includes a `fetch()` polyfill using a `node-fetch` package. Check out the [node-fetch documentation](https://www.npmjs.com/package/node-fetch){:target="_blank"} for usage examples. +### Variable scoping + +When declaring settings variables, make sure to declare them in the function handler rather than globally in your function. This prevents you leaking the settings values across other function instances. + +The handler for insert functions is event-specific, for example, `onTrack()`, `onIdentify()`, and so on. + ### Errors and error handling Segment considers a function's execution successful if it finishes without error. You can `throw` an error to create a failure on purpose. Use these errors to validate event data before processing it to ensure the function works as expected. @@ -176,8 +182,7 @@ async function onIdentify(event) { ``` If you don't supply a function for an event type, Segment throws an `EventNotSupported` error by default. - -You can read more about [error handling](#destination-insert-functions-logs-and-errors) below. +See [errors and error handling](#errors-and-error-handling) for more information on supported error types and how to troubleshoot them. ## Runtime and dependencies diff --git a/src/connections/reverse-etl/reverse-etl-source-setup-guides/databricks-setup.md b/src/connections/reverse-etl/reverse-etl-source-setup-guides/databricks-setup.md index c47619e20a..88ffe7ce51 100644 --- a/src/connections/reverse-etl/reverse-etl-source-setup-guides/databricks-setup.md +++ b/src/connections/reverse-etl/reverse-etl-source-setup-guides/databricks-setup.md @@ -12,24 +12,26 @@ At a high level, when you set up Databricks for Reverse ETL, the configured serv ## Required permissions * Make sure the service principal you use to connect to Segment has permissions to use that warehouse. In the Databricks console go to **SQL warehouses** and select the warehouse you're using. Navigate to **Overview > Permissions** and make sure the service principal you use to connect to Segment has *can use* permissions. +Note the Service Principal UUID from the [User Management Page](https://accounts.cloud.databricks.com/user-management/serviceprincipals/){:target="_blank”} (under Service Principals) for the following SQL operations. + * To grant access to read data from the tables used in the model query, run: ``` - GRANT USAGE ON SCHEMA TO ``; - GRANT SELECT, READ_METADATA ON SCHEMA TO ``; + GRANT USAGE ON SCHEMA TO ``; + GRANT SELECT, READ_METADATA ON SCHEMA TO ``; ``` * To grant Segment access to create a schema to keep track of the running syncs, run: ``` - GRANT CREATE on catalog TO ``; + GRANT CREATE on catalog TO ``; ``` * If you want to create the schema yourself instead and then give Segment access to it, run: ``` CREATE SCHEMA IF NOT EXISTS __segment_reverse_etl; - GRANT ALL PRIVILEGES ON SCHEMA __segment_reverse_etl TO ``; + GRANT ALL PRIVILEGES ON SCHEMA __segment_reverse_etl TO ``; ``` ## Set up guide diff --git a/src/connections/reverse-etl/reverse-etl-source-setup-guides/postgres-setup.md b/src/connections/reverse-etl/reverse-etl-source-setup-guides/postgres-setup.md index 42fe99565d..2a6689f0a8 100644 --- a/src/connections/reverse-etl/reverse-etl-source-setup-guides/postgres-setup.md +++ b/src/connections/reverse-etl/reverse-etl-source-setup-guides/postgres-setup.md @@ -31,6 +31,15 @@ To set up Postgres with Reverse ETL: -- allows the "segment" user to create new schemas on the specified database. (this is the name you chose when provisioning your cluster) GRANT CREATE ON DATABASE "" TO "segment"; + + -- create Segment schema + CREATE SCHEMA __segment_reverse_etl; + + -- Allow user to use the Segment schema + GRANT USAGE ON SCHEMA __segment_reverse_etl TO segment; + + -- Grant all privileges on all existing tables in the Segment schema + GRANT ALL PRIVILEGES ON ALL TABLES IN SCHEMA __segment_reverse_etl TO segment; ``` 4. Make sure the user has correct access permissions to the database. 5. Follow the steps listed in the [Add a source](/docs/connections/reverse-etl/setup/#step-1-add-a-source) section to finish adding Postgres as a source. diff --git a/src/connections/reverse-etl/reverse-etl-source-setup-guides/redshift-setup.md b/src/connections/reverse-etl/reverse-etl-source-setup-guides/redshift-setup.md index 6ae2d4bdc0..c32f6f6aca 100644 --- a/src/connections/reverse-etl/reverse-etl-source-setup-guides/redshift-setup.md +++ b/src/connections/reverse-etl/reverse-etl-source-setup-guides/redshift-setup.md @@ -15,12 +15,21 @@ To set up Redshift with Reverse ETL: 2. Follow the [networking instructions](/docs/connections/storage/catalog/redshift/#networking) to configure the correct network and security settings. 3. Run the SQL commands below to create a user named `segment`. - ```ts + ```sql -- create a user named "segment" that Segment will use when connecting to your Redshift cluster. CREATE USER segment PASSWORD ''; -- allows the "segment" user to create new schemas on the specified database. (this is the name you chose when provisioning your cluster) GRANT CREATE ON DATABASE "" TO "segment"; + + -- create Segment schema + CREATE SCHEMA __segment_reverse_etl; + + -- Allow user to use the Segment schema + GRANT USAGE ON SCHEMA __segment_reverse_etl TO segment; + + -- Grant all privileges on all current tables in the Segment schema + GRANT ALL PRIVILEGES ON ALL TABLES IN SCHEMA __segment_reverse_etl TO segment; ``` 4. Follow the steps listed in the [Add a source](/docs/connections/reverse-etl/setup/#step-1-add-a-source) section to finish adding Redshift as your source. diff --git a/src/connections/reverse-etl/reverse-etl-source-setup-guides/snowflake-setup.md b/src/connections/reverse-etl/reverse-etl-source-setup-guides/snowflake-setup.md index 697b375900..2768240b68 100644 --- a/src/connections/reverse-etl/reverse-etl-source-setup-guides/snowflake-setup.md +++ b/src/connections/reverse-etl/reverse-etl-source-setup-guides/snowflake-setup.md @@ -55,6 +55,7 @@ Follow the instructions below to set up the Segment Snowflake connector. Segment -- database access GRANT USAGE ON DATABASE segment_reverse_etl TO ROLE segment_reverse_etl; GRANT CREATE SCHEMA ON DATABASE segment_reverse_etl TO ROLE segment_reverse_etl; + GRANT CREATE TABLE ON SCHEMA __segment_reverse_etl TO ROLE segment_reverse_etl; ``` 6. Enter and run one of the following code snippets below to create the user Segment uses to run queries. For added security, Segment recommends creating a user that authenticates using a key pair. diff --git a/src/engage/index.md b/src/engage/index.md index b15e7f45d4..f0cdd3d93a 100644 --- a/src/engage/index.md +++ b/src/engage/index.md @@ -22,56 +22,9 @@ Add detail to user profiles with new traits and use them to power personalized m - [**Predictions**:](/docs/unify/traits/predictions/) Predict the likelihood that users will perform custom events tracked in Segment, like LTV, churn, and purchase. #### Build Audiences -Create lists of users or accounts that match specific criteria. For example, after creating an `inactive accounts` audience that lists paid accounts with no logins in 60 days, you can push the audience to your analytics tools or send an SMS, email, or WhatsApp campaign with Engage Channels. Learn more about [Engage audiences](/docs/engage/audiences/). +Create lists of users or accounts that match specific criteria. For example, after creating an `inactive accounts` audience that lists paid accounts with no logins in 60 days, you can push the audience to your analytics tools or send an SMS, email, or WhatsApp campaign with Engage Channels. Learn more about [Engage audiences](/docs/engage/audiences/). #### Sync audiences to downstream tools Once you create your Computed Traits and Audiences, Engage sends them to your Segment Destinations in just a few clicks. You can use these Traits and Audiences to personalize messages across channels, optimize ad spend, and improve targeting. You can also use the [Profile API](/docs/unify/profile-api) to build in-app and onsite personalization. Learn more about [using Engage data](/docs/engage/using-engage-data/) and the [Profile API](/docs/unify/profile-api). -{% include components/reference-button.html href="https://segment.com/customers/drift/" icon="personas.svg" title="Personalizing marketing campaigns" description="Marketing teams use Engage to run real-time multi-channel marketing campaigns based off specific user attributes they've computed in Engage. Read about how Drift used Engage to increase prospect engagement by 150% in two months." %} - -## Market to customers with Engage Premier and Channels - -To send email, SMS, and WhatsApp campaigns with Engage Channels, you'll connect a [Twilio messaging service](https://support.twilio.com/hc/en-us/articles/223181308-Getting-started-with-Messaging-Services){:target="blank"}, [SendGrid subuser account](https://docs.sendgrid.com/ui/account-and-settings/subusers#create-a-subuser){:target="blank"}, and [WhatsApp messaging service](https://www.twilio.com/docs/whatsapp/self-sign-up){:target="blank"} to your Segment Engage space. Use existing accounts, or create new ones. - -View the [onboarding steps](/docs/engage/onboarding/) for more on how to connect Twilio and SendGrid accounts. - -#### Send email, SMS, and WhatsApp messages in Journeys - -Use Engage to build email, SMS, and WhatsApp campaigns within [Journeys](/docs/engage/journeys/). Send campaigns to [subscribed users](#user-subscriptions) based on event behavior and profile traits. With [message analytics](#message-analytics), you can track the performance of your campaigns. - -- **Send Email**: [Build email campaigns](/docs/engage/campaigns/email-campaigns/) with existing templates, or create a new email template within Journeys. Before you send the email, test the template and set [conversion goals](#conversion-goals). - -- **Send SMS messages**: [Build SMS campaigns](/docs/engage/campaigns/sms-campaigns/) to message users in real-time as a step in a Journey. For example, create an abandoned cart campaign that texts users a reminder to complete their purchase, along with a promo code. Add [merge tags](#personalize-with-merge-tags) and set conversion goals. - -- **Send WhatsApp messages**: [Build WhatsApp campaigns](/docs/engage/campaigns/whatsapp-campaigns) that deliver messages to your customers on the world's most used messaging app. - -To learn more, visit the [CSV Uploader](/docs/engage/profiles/csv-upload/) documentation. - -#### Build Email, SMS, and WhatsApp message templates - -Build personalized [email](/docs/engage/content/email/template/), [SMS](/docs/engage/content/sms/template), and [WhatsApp](/docs/engage/content/whatsapp) templates in Twilio Engage for use in your campaigns. Design email templates with a WYSIWYG [Drag and Drop Editor](/docs/engage/content/email/editor/) or the [HTML Editor](/docs/engage/content/email/html-editor/). Engage saves the templates for you to preview, edit, and reuse throughout Journeys. - -#### Personalize with merge tags -Insert real-time user profile traits from merge tags to personalize each message. For example, address recipients by name or highlight new products from a user's favorite brand. - -#### CSV Uploader -Use the CSV uploader to add or update user profiles and [subscription states](/docs/engage/user-subscriptions/). To learn more, visit the [CSV Uploader](/docs/engage/profiles/csv-upload/) documentation. - -#### User subscriptions - -Set user subscription states in two ways: -- [Upload a CSV file](/docs/engage/profiles/csv-upload/) with lists of users along with their phone, email, and WhatsApp subscription states. -- Programmatically with Segment's [Public API](https://api.segmentapis.com/docs/spaces/#replace-messaging-subscriptions-in-spaces){:target="blank"} - -Use Engage to add subscription states to user email addresses and phone numbers. Subscription states help determine which users you can send campaigns to in Engage. You can set user subscription states with a [CSV file upload](/docs/engage/profiles/csv-upload/), or programmatically with Segment's [Public API](https://api.segmentapis.com/docs/spaces/#replace-messaging-subscriptions-in-spaces){:target="blank"}. - -#### Message Analytics -With analytics in Engage, you can monitor real-time conversion data. Track message performance and customer interaction beyond clicks and opens. Use campaign dashboards to view events such as `Email Delivered`, `Unsubscribed`, `Spam Reported`, and more. - -#### Conversion Goals - -For each message step in a Journey, you can set conversion conditions with events and properties in your Segment space. Then, define a duration after message delivery to track goals. - -For example, track users who perform the event **Order Completed** with a promo code that you send them. - -Visit [Message Analytics](/docs/engage/analytics/) to learn more. +{% include components/reference-button.html href="https://segment.com/customers/drift/" icon="personas.svg" title="Personalizing marketing campaigns" description="Marketing teams use Engage to run real-time multi-channel marketing campaigns based off specific user attributes they've computed in Engage. Read about how Drift used Engage to increase prospect engagement by 150% in two months." %} \ No newline at end of file diff --git a/src/privacy/data-retention-policy.md b/src/privacy/data-retention-policy.md index 0a71bc5808..3721bc4558 100644 --- a/src/privacy/data-retention-policy.md +++ b/src/privacy/data-retention-policy.md @@ -44,9 +44,9 @@ Segment recommends keeping your data for at least 30 days to enable [replays](/d To change your data retention settings, open Segment and navigate to **Privacy > Settings > Data Retention**. -### Workspace Default Archive Retention Period +### Workspace default archive retention period -Select the default retention period for the workspace in this setting. This value applies to all sources in the workspace, unless overridden in the [Source-Level Archive Retention Periods](#source-level-archive-retention-periods) setting. +Select the default retention period for the workspace in this setting. This value applies to all sources in the workspace. - 14 days - 30 days @@ -55,21 +55,6 @@ Select the default retention period for the workspace in this setting. This valu - 365 days - 3 years (the default setting starting July 15, 2025) - Unlimited (deprecated July 15, 2025) - -### Source-Level Archive Retention Periods - -> warning "Source-Level Archive Retention Periods will be deprecated on April 15, 2025" -> After April 15, you will no longer be able to override your workspace's default retention period on a source-by-source basis. - -Override the workspace default retention period on a per-source level. - -You can select from the following Archive Retention time periods: -- Default (This is the default value you set in the [Workspace Default Archive Retention Period](#workspace-default-archive-retention-period) setting) -- 14 days -- 30 days -- 90 days -- 180 days -- 365 days ### What data is impacted? @@ -145,4 +130,4 @@ Segment unrecoverably deletes the workspace after 30 days of inactivity, unless ### Data deletion delays -When data reaches the end of its retention period, deletion is scheduled in accordance with Segment’s data retention policy. While Segment aims to complete the deletion process promptly, there may be occasional delays due to processing times or technical constraints. Segment is committed to initiating data deletions as soon as possible and strives to complete deletions within 7 days of the scheduled date. \ No newline at end of file +When data reaches the end of its retention period, deletion is scheduled in accordance with Segment’s data retention policy. While Segment aims to complete the deletion process promptly, there may be occasional delays due to processing times or technical constraints. Segment is committed to initiating data deletions as soon as possible and strives to complete deletions within 7 days of the scheduled date. diff --git a/src/segment-app/extensions/git.md b/src/segment-app/extensions/git.md index 97b19156e4..5dae126d31 100644 --- a/src/segment-app/extensions/git.md +++ b/src/segment-app/extensions/git.md @@ -6,7 +6,10 @@ Segment's Git extension lets you manage versioning by syncing changes you make i Git Sync supports synchronization from Segment to Git. When you sync data from Segment to Git, you capture the current state of your workspace through a full sync and includes all new records and changes for supported resources. -You can use [bidirectional sync](#bidirectional-sync) to sync data from Git to Segment. After you enable bidirectional sync, Segment automatically listens for pull requests in your repository and manages all related workspace changes. +You can use [bidirectional sync](#bidirectional-sync) to sync data from Git to Segment. After you enable bidirectional sync, Segment automatically listens for pull requests in your repository and manages all related workspace changes. + +> info "Bidirectional sync is in Private Beta" +> Bidirectional sync is in private beta, and Segment is actively working on this feature. Some functionality may change before it becomes generally available. ## Set up Git Sync @@ -88,6 +91,9 @@ For more information on using Terraform, visit [Terraform's documentation](https Bidirectional sync builds on top of the Git Sync extension and lets you manage your Segment workspace directly in GitHub. After you configure and enable bidirectional sync, Segment automatically listens for pull requests in your repository and manages all related workspace changes. Segment only applies changes when you comment `segment apply` on pull requests that can be successfully merged. +> info "Bidirectional sync is in Private Beta" +> Bidirectional sync is in private beta, and Segment is actively working on this feature. Some functionality may change before it becomes generally available. + Bidirectional sync only supports: - Explicit values ([secrets](#use-secrets-with-bidirectional-sync) require additional configuration) - [Segment resources compatible with Git sync](#working-with-git-sync) diff --git a/src/unify/data-graph/index.md b/src/unify/data-graph/index.md index 2061cb55e7..4860be27e1 100644 --- a/src/unify/data-graph/index.md +++ b/src/unify/data-graph/index.md @@ -13,9 +13,6 @@ The Data Graph acts as a semantic layer that allows businesses to define relatio ## Prerequisites -> info "Why you need both materialized and unmaterialized tables" -> Segment recommends using materialized views for Profiles Sync to optimize performance and reduce query costs with Linked Audiences. However, due to schema inference requirements, you still need to select the matching **unmaterialized tables** as well. Segment relies on the unmaterialized tables during setup, even if they’re not used when queries run. - To use the Data Graph, you'll need the following: - A supported data warehouse with the appropriate Data Graph permissions @@ -354,13 +351,25 @@ For many:many relationships, define the join on between the two entity tables wi **Junction table spec** -| Parameters | Definition | -| --------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | +| Parameters |Definition | +| --------------- | --------------------------------- | | `table_ref` | Defines the fully qualified table reference to the join table: `[database name].[schema name].[table name]`. Segment flexibly supports tables, views and materialized views | | `primary_key` | The unique identifier for the given table. Must be a column with unique values per row | | `left_join_on` | Define the relationship between the left entity table and the junction table: `[left entity slug].[column name] = [junction table column name]`. Note that schema and table are implied within the junction table column name, so you do not need to define it again | | `right_join_on` | Define the relationship between the junction table and the right entity table: `[junction table column name] = [right entity slug].[column name]`. Note that schema and table are implied within the junction table column name, so you do not need to define it again | + +When you define a many-to-many relationship using a junction table, `left_join_on` and `right_join_on` tell Data Graph how to connect each entity to the junction table: + +* Use `left_join_on` to specify which column in the junction table links to the parent (left) entity. + +* Use `right_join_on` to specify which column links to the child (right) entity. + +These fields define the join conditions, but they don’t control how the join is executed. Data Graph always performs inner joins, even if you specify a `left_join_on`. + +If you need behavior similar to a left join (like including unmatched rows), create a view in your warehouse with the logic you’re targeting and reference that view as an entity in your graph. + + **Example:** ```python diff --git a/src/unify/data-graph/linked-events.md b/src/unify/data-graph/linked-events.md index 407b8b303f..ea32cb189e 100644 --- a/src/unify/data-graph/linked-events.md +++ b/src/unify/data-graph/linked-events.md @@ -159,7 +159,7 @@ To configure your sync schedule: 3. Click **Edit** next to **Sync schedule**. 4. Select the **Schedule type**. You can choose from: * **Manual**: Trigger the sync manually or with Segment's API. - * **Interval**: Sync based on a by-the minute, hourly, or daily cycle. For example, once every 2 hours. + * **Interval**: Sync at predefined intervals: 15 min, 30 min, 1 hour, 2 hours, 4 hours, 6 hours, 8 hours, 12 hours, or 1 day * **Day and time**: Sync at specific times on selected days of the week. For example, Mondays at 2:00PM. ### Add entities @@ -213,7 +213,7 @@ To use Linked Events, be sure that you have proper permissions for the Data Ware #### How often do syncs occur? -Segment currently syncs once every hour. +You can configure your syncs to occur at predefined intervals: 15 min, 30 min, 1 hour, 2 hours, 4 hours, 6 hours, 8 hours, 12 hours, or 1 day. See the section on [configuring the sync schedule](#configure-the-sync-schedule) to learn more. #### Which Destinations does Linked Events support?