From 67a73c9c22c4c711e45c64450a8af5ed47dafdf8 Mon Sep 17 00:00:00 2001 From: Ryan Curtis Date: Thu, 10 Oct 2024 13:59:30 -0400 Subject: [PATCH 1/2] remove font weight hover effect for nav bar (#1525) --- pages/overrides.scss | 63 ++++++-------------------------------------- 1 file changed, 8 insertions(+), 55 deletions(-) diff --git a/pages/overrides.scss b/pages/overrides.scss index add5d31e2d..77ada110c2 100644 --- a/pages/overrides.scss +++ b/pages/overrides.scss @@ -91,13 +91,6 @@ $border-radius-lg: 100px; font-size: 14px; line-height: 20px; padding: 8px; - - //hover effect - & span:hover { - font-weight: 700; - // color: colors.$purple100; - // text-decoration: underline !important; - } } .top-nav-search-bar { @@ -659,30 +652,10 @@ article img { // bold the parent links, first two are desktop version, second two are mobile version // extra specificity is required because we want to distinguish between the top level and nested files - .nextra-scrollbar - > ._transform-gpu - > ul - > li - > a, - .nextra-scrollbar - > ._transform-gpu - > ul - > li - > button, - .nextra-scrollbar - > ul - > li.open - > ._transform-gpu - > ul - > li - > a, - .nextra-scrollbar - > ul - > li.open - > ._transform-gpu - > ul - > li - > button { + .nextra-scrollbar > ._transform-gpu > ul > li > a, + .nextra-scrollbar > ._transform-gpu > ul > li > button, + .nextra-scrollbar > ul > li.open > ._transform-gpu > ul > li > a, + .nextra-scrollbar > ul > li.open > ._transform-gpu > ul > li > button { font-weight: 500; } @@ -1344,30 +1317,10 @@ article img { // bold the parent links, first two are desktop version, second two are mobile version // extra specificity is required because we want to distinguish between the top level and nested files - .nextra-scrollbar - > ._transform-gpu - > ul - > li - > a, - .nextra-scrollbar - > ._transform-gpu - > ul - > li - > button, - .nextra-scrollbar - > ul - > li.open - > ._transform-gpu - > ul - > li - > a, - .nextra-scrollbar - > ul - > li.open - > ._transform-gpu - > ul - > li - > button { + .nextra-scrollbar > ._transform-gpu > ul > li > a, + .nextra-scrollbar > ._transform-gpu > ul > li > button, + .nextra-scrollbar > ul > li.open > ._transform-gpu > ul > li > a, + .nextra-scrollbar > ul > li.open > ._transform-gpu > ul > li > button { font-weight: 500; } From f96445757c4f8708d084e0f03b7da2d0959ddaa7 Mon Sep 17 00:00:00 2001 From: Tiffany Qi Date: Thu, 10 Oct 2024 11:54:42 -0700 Subject: [PATCH 2/2] Fix some broken links (#1524) * fix some broken links * a couple more --- ...2023-11-07-lookup-tables-warehouse-connectors.mdx | 4 ++-- pages/docs/access-security/single-sign-on.mdx | 2 +- pages/docs/access-security/single-sign-on/azure.md | 2 +- pages/docs/cohort-sync/integrations.mdx | 4 ++-- .../data-pipelines/old-pipelines/integrations.mdx | 2 +- .../integrations/schematized-aws-pipeline.md | 12 ++++++------ .../integrations/schematized-azure-pipeline.md | 4 ++-- .../integrations/schematized-bigquery-pipeline.md | 6 +++--- .../integrations/schematized-gcs-pipeline.md | 2 +- .../integrations/schematized-snowflake-pipeline.md | 4 ++-- pages/docs/data-structure/user-profiles.mdx | 4 ++-- pages/docs/migration/google-analytics.md | 2 +- pages/docs/quickstart/connect-your-data.mdx | 2 +- pages/docs/quickstart/identify-users.mdx | 2 +- pages/docs/quickstart/track-events.mdx | 2 +- .../migrating-to-simplified-id-merge-system.md | 6 +++--- redirects/developer-mixpanel-com.txt | 4 ++-- redirects/local.txt | 6 +++--- 18 files changed, 35 insertions(+), 35 deletions(-) diff --git a/pages/changelogs/2023-11-07-lookup-tables-warehouse-connectors.mdx b/pages/changelogs/2023-11-07-lookup-tables-warehouse-connectors.mdx index 1f28a1143d..1b6d13c79e 100644 --- a/pages/changelogs/2023-11-07-lookup-tables-warehouse-connectors.mdx +++ b/pages/changelogs/2023-11-07-lookup-tables-warehouse-connectors.mdx @@ -6,6 +6,6 @@ createdAt: "2023-11-07T14:59:02.165Z" updatedAt: "2023-11-07T14:59:02.165Z" date: "2023-11-07" --- -Today, all customers can use Warehouse Connectors to import Lookup Tables from their warehouse. For a list of supported warehouses, check out this [list](https://docs.mixpanel.com/docs/tracking-methods/data-warehouse/overview). This capability allows you to periodically enrich Mixpanel with data from your source-of-truth. +Today, all customers can use Warehouse Connectors to import Lookup Tables from their warehouse. For a list of supported warehouses, check out this [list](/docs/tracking-methods/data-warehouse). This capability allows you to periodically enrich Mixpanel with data from your source-of-truth. -For more information, check out our [docs](https://docs.mixpanel.com/docs/tracking-methods/data-warehouse/sending-lookup-table). +For more information, check out our [docs](/docs/tracking-methods/data-warehouse#lookup-tables). diff --git a/pages/docs/access-security/single-sign-on.mdx b/pages/docs/access-security/single-sign-on.mdx index 3d8e7e6a60..3a5ba77081 100644 --- a/pages/docs/access-security/single-sign-on.mdx +++ b/pages/docs/access-security/single-sign-on.mdx @@ -21,7 +21,7 @@ Claiming a domain will add security to an SSO implementation by allowing only me To claim a domain, add a TXT record to your domain's DNS records with a verification key provided by Mixpanel. The verification key is available after you submit to claim a domain in your Organization Settings. - A single domain can be claimed by only one single Mixpanel organization. If you have multiple Mixpanel organizations with separate billing accounts but need to share SSO settings and email domains for login, please see the [documentation for the Shared SSO beta](docs/access-security/single-sign-on/shared-sso) to see how you can share those settings with an "Admin" Organization. + A single domain can be claimed by only one single Mixpanel organization. If you have multiple Mixpanel organizations with separate billing accounts but need to share SSO settings and email domains for login, please see the [documentation for the Shared SSO beta](/docs/access-security/single-sign-on/shared-sso) to see how you can share those settings with an "Admin" Organization. ### Generate Verification Key by Claiming Domain diff --git a/pages/docs/access-security/single-sign-on/azure.md b/pages/docs/access-security/single-sign-on/azure.md index cccb593bc5..c1c3506092 100644 --- a/pages/docs/access-security/single-sign-on/azure.md +++ b/pages/docs/access-security/single-sign-on/azure.md @@ -3,7 +3,7 @@ ## Overview -You can set up Microsoft Azure Single Sign-On to use with your Mixpanel account. Before using this document, read the [general Mixpanel SSO set-up instructions]((https://docs.microsoft.com/en-us/azure/active-directory/saas-apps/mixpanel-provisioning-tutorial)) as well as [Azure's documentation on setting up a new application](https://docs.microsoft.com/en-us/azure/active-directory/saas-apps/mixpanel-provisioning-tutorial). +You can set up Microsoft Azure Single Sign-On to use with your Mixpanel account. Before using this document, read the [general Mixpanel SSO set-up instructions](/docs/access-security/single-sign-on) as well as [Azure's documentation on setting up a new application](https://docs.microsoft.com/en-us/azure/active-directory/saas-apps/mixpanel-provisioning-tutorial). ### Add Mixpanel as a New Application diff --git a/pages/docs/cohort-sync/integrations.mdx b/pages/docs/cohort-sync/integrations.mdx index 2deb384e41..9972c51662 100644 --- a/pages/docs/cohort-sync/integrations.mdx +++ b/pages/docs/cohort-sync/integrations.mdx @@ -6,7 +6,7 @@ import { Callout, Cards } from 'nextra/components' This page highlights integrations that sends Mixpanel Cohorts to external destinations. Learn more about [integrations that imports data from other destinations into Mixpanel](/docs/tracking-methods/integrations) and [integrations that exports data from Mixpanel to data warehouses](/docs/data-pipelines/integrations/). -These integrations allow you to send your [Cohorts](docs/users/cohorts) to other destinations for audience targeting. You may review our complete integrations directory [here](https://mixpanel.com/partners/integrations). If you don't see what you're looking for [reach out to us](https://mixpanel.com/get-support). +These integrations allow you to send your [Cohorts](/docs/users/cohorts) to other destinations for audience targeting. You may review our complete integrations directory [here](https://mixpanel.com/partners/integrations). If you don't see what you're looking for [reach out to us](https://mixpanel.com/get-support). ## Featured Integrations @@ -28,7 +28,7 @@ These integrations allow you to send your [Cohorts](docs/users/cohorts) to other - + diff --git a/pages/docs/data-pipelines/old-pipelines/integrations.mdx b/pages/docs/data-pipelines/old-pipelines/integrations.mdx index 65462a57c8..9565f1d5cb 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations.mdx +++ b/pages/docs/data-pipelines/old-pipelines/integrations.mdx @@ -6,7 +6,7 @@ import { Callout, Cards } from "nextra/components"; This page highlights integrations that send Mixpanel data to data warehouses. Learn more about [integrations that import data from other destinations into Mixpanel](/docs/tracking-methods/integrations) and [integrations that export - cohorts for audience targeting](docs/cohort-sync/integrations). + cohorts for audience targeting](/docs/cohort-sync/integrations). These integrations allow you to send your [raw data](/docs/data-pipelines#raw) and [schematized data](/docs/data-pipelines#schematized) to your cloud storage bucket or data warehouse. You may review our complete integrations directory [here](https://mixpanel.com/partners/integrations). If you don't see what you're looking for [reach out to us](https://mixpanel.com/get-support). diff --git a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline.md b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline.md index 7671dac217..b88985db85 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline.md +++ b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline.md @@ -1,11 +1,11 @@ # Schematized AWS Pipeline -Mixpanel's [Schematized Export Pipeline](/docs/data-pipelines/schematized-export-pipeline) lets you export your Mixpanel data directly into an S3 bucket, allowing the use of Glue to query it. To set up the Mixpanel AWS pipeline, you must configure AWS to receive the exported data, then [create a pipeline](https://developer.mixpanel.com/reference/create-warehouse-pipeline) to export the data. +Mixpanel's [Schematized Export Pipeline](/docs/data-pipelines/old-pipelines/schematized-export-pipeline) lets you export your Mixpanel data directly into an S3 bucket, allowing the use of Glue to query it. To set up the Mixpanel AWS pipeline, you must configure AWS to receive the exported data, then [create a pipeline](https://developer.mixpanel.com/reference/create-warehouse-pipeline) to export the data. ## Design ![image](/230698348-abb2656e-fe2a-4d9c-ad61-8f80793e9c07.png) -Mixpanel applies [transformation rules](/docs/data-pipelines/schematized-export-pipeline#transformation-rules) to make the data compatible with data warehouses and then transfers the transformed data to your S3 bucket. You can then choose to use a Glue crawler to create the schema out of the transformed data or let Mixpanel to directly create the schema in your glue database. Having the data and the glue schema in place, you can use SQL with multiple AWS products, including [Amazon Athena](https://aws.amazon.com/athena/) and [Redshift Spectrum](https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html), to query the imported data. +Mixpanel applies [transformation rules](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules) to make the data compatible with data warehouses and then transfers the transformed data to your S3 bucket. You can then choose to use a Glue crawler to create the schema out of the transformed data or let Mixpanel to directly create the schema in your glue database. Having the data and the glue schema in place, you can use SQL with multiple AWS products, including [Amazon Athena](https://aws.amazon.com/athena/) and [Redshift Spectrum](https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html), to query the imported data. We recommend the use of [Spectrum](https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html) and [Glue](https://aws.amazon.com/glue/) to query the imported data. @@ -164,7 +164,7 @@ Note: If you use Athena with Glue and want to enable partitioning, you must choo ##### Configuring Glue for Mixpanel Direct Export -Mixpanel can write and update a schema in your Glue instance as soon as new data is available. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/schematized-export-pipeline#schema). To set this up: +Mixpanel can write and update a schema in your Glue instance as soon as new data is available. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#schema). To set this up: 1. Create a Glue database. * In the AWS console, go to "Glue". (Make sure you are in the same region as your S3 exported data). @@ -265,7 +265,7 @@ This convenience, however, can also pose issues, such as: * The time column is interpreted as an integer. As a result, it is more difficult to perform timestamp operations. * When using the one table per event schema option, Glue crawlers can merge data from multiple events in one table based on similarity. -To avoid these issues, Mixpanel can write and update a schema in your Glue instance as soon as new data is available. Follow [these instructions](/docs/data-pipelines/integrations/amazon-s3#section-configuring-glue-for-mixpanel-direct-export) to enable Mixpanel to write your data catalog to AWS Glue. +To avoid these issues, Mixpanel can write and update a schema in your Glue instance as soon as new data is available. Follow [these instructions](/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline#configuring-glue-for-mixpanel-direct-export) to enable Mixpanel to write your data catalog to AWS Glue. To use crawlers, you must point the crawler to the top level folder with your Mixpanel project ID. @@ -292,7 +292,7 @@ You only need to connect Redshift to Glue once. As the Mixpanel pipeline exports You can also join data stored in Redshift with Mixpanel data available in S3 through the external schema. ## Queries -You can query data with a single table schema or with a multiple table schema in Redshift Spectrum. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/schematized-export-pipeline#schema). +You can query data with a single table schema or with a multiple table schema in Redshift Spectrum. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#schema). To query a single table schema, use this snippet. ```sql @@ -308,7 +308,7 @@ FROM mixpanel.mp_master_event WHERE mp_event_name = “” ``` -`CLEANED_EVENT_NAME` is the transformed event name based on [transformation rules](/docs/data-pipelines/schematized-export-pipeline#transformation-rules). +`CLEANED_EVENT_NAME` is the transformed event name based on [transformation rules](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules). #### Nested and Repeated Fields diff --git a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-azure-pipeline.md b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-azure-pipeline.md index 45edba2a64..5f1c5afac3 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-azure-pipeline.md +++ b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-azure-pipeline.md @@ -1,12 +1,12 @@ # Schematized Azure Pipeline -Mixpanel’s Azure Blob Export is one type of [Schematized Export Pipeline](/docs/data-pipelines/schematized-export-pipeline) that exports Mixpanel events and people data and imports them into an Azure Blob storage instance. +Mixpanel’s Azure Blob Export is one type of [Schematized Export Pipeline](/docs/data-pipelines/old-pipelines/schematized-export-pipeline) that exports Mixpanel events and people data and imports them into an Azure Blob storage instance. Azure Export enables the use of SQL to query your event data, either for expressing new types of queries or for joining against other data already present in your data environment. ## Integration Set Up -In order to export data to Azure Blob Storage, you must first give Mixpanel permission to write to Azure. Next you can [create a pipeline](https://developer.mixpanel.com/reference/create-warehouse-pipeline) to export the data. Mixpanel then uploads the data to Azure Blob Storage on a recurring basis. The data is a slightly transformed version of the raw data stored in Mixpanel. Please see [transformation rules](/docs/data-pipelines/schematized-export-pipeline#transformation-rules) to understand how the data is transformed. +In order to export data to Azure Blob Storage, you must first give Mixpanel permission to write to Azure. Next you can [create a pipeline](https://developer.mixpanel.com/reference/create-warehouse-pipeline) to export the data. Mixpanel then uploads the data to Azure Blob Storage on a recurring basis. The data is a slightly transformed version of the raw data stored in Mixpanel. Please see [transformation rules](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules) to understand how the data is transformed. ## Edit Mixpanel Permission diff --git a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-bigquery-pipeline.md b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-bigquery-pipeline.md index 8a3f2dbc69..ad617d2df3 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-bigquery-pipeline.md +++ b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-bigquery-pipeline.md @@ -10,7 +10,7 @@ There are currently two ways to export mixpanel data into big. ![image](/230698685-c02cb9a1-d66f-42a7-8063-8e78b79e7b1f.png) -As part of the export pipeline, a new dataset `mixpanel_nessie_day_partitioned_` will be created if the customer chose to export into Mixpanel managed BigQuery or a dataset should be created with appropriate permissions on customer-managed BigQuery and provide us with a name. We also apply [transformation rules](/docs/data-pipelines/schematized-export-pipeline#transformation-rules) to make the data compatible with data warehouses. +As part of the export pipeline, a new dataset `mixpanel_nessie_day_partitioned_` will be created if the customer chose to export into Mixpanel managed BigQuery or a dataset should be created with appropriate permissions on customer-managed BigQuery and provide us with a name. We also apply [transformation rules](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules) to make the data compatible with data warehouses. For user profile and identity mappings tables, we create a new table with a random suffix every time and then will update the `mp_people` and `mp_identity_mappings` views accordingly to use the latest table. You should always use the views and should refrain from using the actual tables as we don't delete the old tables immediately and you may be using an old table. @@ -22,7 +22,7 @@ The data in the tables is partitioned based on [`_PARTITIONTIME` pseudo column]( Note: TIMEPARITIONING shouldn't be updated on the table. It will fail your export jobs. Create a new table/view from this table for custom partitioning. ## Queries -You can query data with a single table schema or with a multiple table schema in BigQuery. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/schematized-export-pipeline#schema). +You can query data with a single table schema or with a multiple table schema in BigQuery. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#schema). To query a single table schema, use this snippet. ```sql @@ -37,7 +37,7 @@ SELECT count(*) FROM mixpanel_nessie_day_partitioned_. ``` -`CLEANED_EVENT_NAME` is the transformed event name based on [transformation rules](/docs/data-pipelines/schematized-export-pipeline#transformation-rules). +`CLEANED_EVENT_NAME` is the transformed event name based on [transformation rules](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules). #### Getting the number of events in each day You will need this if you suspect the export process is not exporting all the events you want. As the tables are partitions using [`_PARTITIONTIME` pseudo column](https://cloud.google.com/bigquery/docs/querying-partitioned-tables#ingestion-time_partitioned_table_pseudo_columns) and in project timezone, you can use to following query to get the number of events per day in an easy and fast way: diff --git a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-gcs-pipeline.md b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-gcs-pipeline.md index 69f3fd82a4..f64c654a17 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-gcs-pipeline.md +++ b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-gcs-pipeline.md @@ -1,6 +1,6 @@ # Schematized GCS Pipeline -The Google Cloud Storage schema pipeline exports [transformed data](/docs/data-pipelines/schematized-export-pipeline#transformation-rules) to a GCS bucket. This can be useful in case you want to export Mixpanel data into your own BigQuery instance. +The Google Cloud Storage schema pipeline exports [transformed data](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#transformation-rules) to a GCS bucket. This can be useful in case you want to export Mixpanel data into your own BigQuery instance. To set up a schematized export pipeline to Google Cloud Storage (GCS) from Mixpanel, you must configure GCS to receive the exported data, then [create a pipeline](https://developer.mixpanel.com/reference/create-warehouse-pipeline) to export the data. diff --git a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-snowflake-pipeline.md b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-snowflake-pipeline.md index c4aded4c16..e0d19befa4 100644 --- a/pages/docs/data-pipelines/old-pipelines/integrations/schematized-snowflake-pipeline.md +++ b/pages/docs/data-pipelines/old-pipelines/integrations/schematized-snowflake-pipeline.md @@ -6,7 +6,7 @@ This guide describes how Mixpanel data is exported into a [Snowflake](https://do Mixpanel exports data to its own Snowflake account and gives your Snowflake account access to read the data. As a result, you will need a Snowflake account to be able to use the exported data. For more information on how to use your shared data, see [Sharing Data in Snowflake](https://docs.snowflake.net/manuals/user-guide-data-share.html). -For Snowflake export, we first load the data into a single-column raw (VARIANT type) data table. This data will be a [transformed version](/docs/data-pipelines/schematized-export-pipeline#section-transformation-rules) of the raw data stored in Mixpanel. Then, we create a view to expose all properties as columns. The view name is the cleaned version of the event name and the raw table name is always the view name suffixed with `_raw`. For example, for `signup` event, you will have a `signup_raw` table and a `signup` view. Please see [Schema](/docs/data-pipelines/schematized-export-pipeline#schema) for general information about the schemas in Schematized Export Pipelines. +For Snowflake export, we first load the data into a single-column raw (VARIANT type) data table. This data will be a [transformed version](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#section-transformation-rules) of the raw data stored in Mixpanel. Then, we create a view to expose all properties as columns. The view name is the cleaned version of the event name and the raw table name is always the view name suffixed with `_raw`. For example, for `signup` event, you will have a `signup_raw` table and a `signup` view. Please see [Schema](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#schema) for general information about the schemas in Schematized Export Pipelines. ## Partitioning @@ -14,7 +14,7 @@ The data in the raw tables is clustered based on `time` column but in project's ## Queries -Mixpanel recommends you place all events into a single table to make querying easier. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/schematized-export-pipeline#schema). +Mixpanel recommends you place all events into a single table to make querying easier. To get more information about the table schemas, please see [Schema](/docs/data-pipelines/old-pipelines/schematized-export-pipeline#schema). A query is a request for data results. You can perform actions on the data, such as combine data from different tables; add, change, or delete table data; and perform calculations. diff --git a/pages/docs/data-structure/user-profiles.mdx b/pages/docs/data-structure/user-profiles.mdx index ec38492a3a..2f3b7bf445 100644 --- a/pages/docs/data-structure/user-profiles.mdx +++ b/pages/docs/data-structure/user-profiles.mdx @@ -30,9 +30,9 @@ For more information about user profiles refer to the documentation on [The Mixp ## Importing Profiles via API -You can create or update User Profiles in similar ways you track events: from our [SDKs](/docs/tracking-methods/sdks/javascript#storing-user-profiles), via our [HTTP Engage API](https://developer.mixpanel.com/reference/profile-set), [Warehouse Connectors](/docs/tracking-methods/data-warehouse/sending-user-profiles), or via our integrations partners. +You can create or update User Profiles in similar ways you track events: from our [SDKs](/docs/tracking-methods/sdks/javascript#storing-user-profiles), via our [HTTP Engage API](https://developer.mixpanel.com/reference/profile-set), [Warehouse Connectors](/docs/tracking-methods/data-warehouse#user-profiles), or via our integrations partners. -Similarly for [Group Profiles](/docs/data-structure/group-analytics#group-profiles), they can be created or updated using our [SDKs](/docs/tracking-methods/sdks/javascript#creating-group-profiles), via our [HTTP Groups API](https://developer.mixpanel.com/reference/group-set-property), [Warehouse Connectors](/docs/tracking-methods/data-warehouse/sending-group-profiles), or via our integration partners. +Similarly for [Group Profiles](/docs/data-structure/group-analytics#group-profiles), they can be created or updated using our [SDKs](/docs/tracking-methods/sdks/javascript#creating-group-profiles), via our [HTTP Groups API](https://developer.mixpanel.com/reference/group-set-property), [Warehouse Connectors](/docs/tracking-methods/data-warehouse#group-profiles), or via our integration partners. We recommend tracking profiles from as close as possible to the source of truth, which is usually your application database or your CRM. One typical approach (especially for [Server-Side Tracking](/docs/tracking-methods/choosing-the-right-method#server-side-tracking)) is to run an hourly or daily script on your servers that pulls the list of profiles from your database and pushes them to Mixpanel. diff --git a/pages/docs/migration/google-analytics.md b/pages/docs/migration/google-analytics.md index c5a7ebea85..7d211ad3ea 100644 --- a/pages/docs/migration/google-analytics.md +++ b/pages/docs/migration/google-analytics.md @@ -92,7 +92,7 @@ Preview the users and associated user properties to ensure the data looks as exp ![image](/ga4_users.png) ##### Post-migration data validation -You can use our [Lexicon](/docs/data-governance/lexicon) or Events page to check that your data has successfully been ingested. However, if your historical events are older than 30 days, they will not show up on Lexicon, Events page or in the event dropdown menu across all reports. In this case, you can leverage our [Insights report](docs/reports/insights) to validate the historical events, by selecting the import time frame and filtering by the following default properties: +You can use our [Lexicon](/docs/data-governance/lexicon) or Events page to check that your data has successfully been ingested. However, if your historical events are older than 30 days, they will not show up on Lexicon, Events page or in the event dropdown menu across all reports. In this case, you can leverage our [Insights report](/docs/reports/insights) to validate the historical events, by selecting the import time frame and filtering by the following default properties: - Warehouse Import ID (tracked as `$warehouse_import_id`) - Warehouse Import Job ID (`$warehouse_import_job_id`) diff --git a/pages/docs/quickstart/connect-your-data.mdx b/pages/docs/quickstart/connect-your-data.mdx index f16dc5b2dd..33df7c2f87 100644 --- a/pages/docs/quickstart/connect-your-data.mdx +++ b/pages/docs/quickstart/connect-your-data.mdx @@ -477,7 +477,7 @@ See our [API reference](https://developer.mixpanel.com/reference/events) for mor **Best Practices for Scale** -You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/integrations/amazon-s3) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. +You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. Here are some other best practices: diff --git a/pages/docs/quickstart/identify-users.mdx b/pages/docs/quickstart/identify-users.mdx index 776ef34043..852d1a821f 100644 --- a/pages/docs/quickstart/identify-users.mdx +++ b/pages/docs/quickstart/identify-users.mdx @@ -227,7 +227,7 @@ print(response.text) **Best Practices for Scale** -You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/integrations/amazon-s3) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. +You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. Here are some other best practices: diff --git a/pages/docs/quickstart/track-events.mdx b/pages/docs/quickstart/track-events.mdx index 373cdae14b..0ad66f5a5e 100644 --- a/pages/docs/quickstart/track-events.mdx +++ b/pages/docs/quickstart/track-events.mdx @@ -268,7 +268,7 @@ See our [API reference](https://developer.mixpanel.com/reference/events) for mor **Best Practices for Scale** -You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/integrations/amazon-s3) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. +You can use this API at scale, for example to backfill historical data into Mixpanel or as part of a high-throughput streaming pipeline. We provide walkthroughs for [Amazon S3](/docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline) and [Google Cloud Storage](/docs/data-pipelines/integrations/google-cloud-storage) to provide a more production-grade example of how to use this API at scale. Here are some other best practices: diff --git a/pages/docs/tracking-methods/id-management/migrating-to-simplified-id-merge-system.md b/pages/docs/tracking-methods/id-management/migrating-to-simplified-id-merge-system.md index fd06754a8b..1fce5b16d7 100644 --- a/pages/docs/tracking-methods/id-management/migrating-to-simplified-id-merge-system.md +++ b/pages/docs/tracking-methods/id-management/migrating-to-simplified-id-merge-system.md @@ -282,7 +282,7 @@ The following guide outlines the steps required to set up the new Mixpanel proje 3. Continue to set up the new project by following the guide [here](/docs/best-practices/project-setup). Configure the new project's settings by referring to your existing project’s settings. Some of the setup tasks may include: - - [Inviting users]([/docs/orgs-and-projects/roles-and-permissions#invite-users-to-a-project) to the project and granting them [Roles and Permissions](/docs/orgs-and-projects/roles-and-permissions#permissions) + - [Inviting users](/docs/orgs-and-projects/roles-and-permissions#invite-users-to-a-project) to the project and granting them [Roles and Permissions](/docs/orgs-and-projects/roles-and-permissions#permissions) - Creating [Teams](/docs/orgs-and-projects/roles-and-permissions#teams) and adding users to them - Adding [group keys](/docs/data-structure/group-analytics#group-keys-in-project-settings) from Group Analytics - Creating [data views](/docs/data-governance/data-views-and-classification#create-data-view) and adding users or teams to them @@ -334,7 +334,7 @@ Update your tech stack with the new project’s token, API secret, and service a 4. Other 3rd-party Platforms integration: - If you are sending events from 3rd-party platforms (i.e. attribution and messaging tools), make sure to update the Mixpanel project token (or API secret) and ensure that 3rd-party events sent to Mixpanel are compatible with Simplified ID Merge. Refer to this [section](docs/tracking-methods/id-management/identifying-users#how-should-i-link-identified-ids-from-3rd-party-systems) for more information. + If you are sending events from 3rd-party platforms (i.e. attribution and messaging tools), make sure to update the Mixpanel project token (or API secret) and ensure that 3rd-party events sent to Mixpanel are compatible with Simplified ID Merge. Refer to this [section](/docs/tracking-methods/id-management/identifying-users#how-should-i-link-identified-ids-from-3rd-party-systems) for more information.
@@ -430,7 +430,7 @@ As part of creating the new Simplfied ID Merge project, you would also need to m Before you move any Board, it's important to note the following: - Duplicate the existing Board and move the new copy into the new project. This would minimise impact where users are still using Boards and reports in the old project. - Any saved cohorts, custom events, custom properties, lookup tables would need to be created first as they don't get automatically moved as part of the Move Board. - - You may need to replicate the permissions for the moved Board should you have very specific [sharing permissions](docs/boards/advanced#sharing) set in the existing project. + - You may need to replicate the permissions for the moved Board should you have very specific [sharing permissions](/docs/boards/advanced#sharing) set in the existing project. - Double check that all reports (especially those that use cohorts, custom events, custom properties, lookup tables) are working properly.
diff --git a/redirects/developer-mixpanel-com.txt b/redirects/developer-mixpanel-com.txt index df5460284d..d41df4996f 100644 --- a/redirects/developer-mixpanel-com.txt +++ b/redirects/developer-mixpanel-com.txt @@ -41,7 +41,7 @@ /developer-docs-redirect/jql-language-reference /docs/reports/apps/jql /developer-docs-redirect/jql-overview /docs/reports/apps/jql /developer-docs-redirect/jql-tutorial-building-an-application /docs/reports/apps/jql -/developer-docs-redirect/mixpanel-amazon-s3-export /docs/data-pipelines/integrations/amazon-s3 +/developer-docs-redirect/mixpanel-amazon-s3-export /docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline /developer-docs-redirect/mixpanel-bigquery-export-design /docs/data-pipelines/integrations/bigquery /developer-docs-redirect/mixpanel-snowflake-export /docs/data-pipelines/integrations/snowflake /developer-docs-redirect/mparticle /docs/tracking-methods/integrations/mparticle @@ -60,7 +60,7 @@ /developer-docs-redirect/ruby /docs/tracking-methods/sdks/ruby /developer-docs-redirect/rudderstack /docs/tracking-methods/integrations/rudderstack /developer-docs-redirect/s3-import /docs/tracking-methods/integrations/amazon-s3 -/developer-docs-redirect/schematized-export-pipeline /docs/data-pipelines/schematized-export-pipeline +/developer-docs-redirect/schematized-export-pipeline /docs/data-pipelines/old-pipelines/schematized-export-pipeline /developer-docs-redirect/segment /docs/tracking-methods/integrations/segment /developer-docs-redirect/server /docs/quickstart/install-mixpanel?sdk=python /developer-docs-redirect/snowplow /docs/tracking-methods/integrations/snowplow diff --git a/redirects/local.txt b/redirects/local.txt index d70e977936..1ee146831a 100644 --- a/redirects/local.txt +++ b/redirects/local.txt @@ -83,10 +83,10 @@ /docs/other-bits/data-pipelines/azure-raw-pipeline /docs/data-pipelines/integrations/azure-raw-pipeline /docs/other-bits/data-pipelines/gcs /docs/data-pipelines/integrations/google-cloud-storage /docs/other-bits/data-pipelines/gcs-raw-pipeline /docs/data-pipelines/integrations/google-cloud-storage-raw-pipeline -/docs/other-bits/data-pipelines/mixpanel-amazon-s3-export /docs/data-pipelines/integrations/amazon-s3 +/docs/other-bits/data-pipelines/mixpanel-amazon-s3-export /docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline /docs/other-bits/data-pipelines/mixpanel-bigquery-export-design /docs/data-pipelines/integrations/bigquery /docs/other-bits/data-pipelines/mixpanel-snowflake-export /docs/data-pipelines/integrations/snowflake -/docs/other-bits/data-pipelines/schematized-export-pipeline /docs/data-pipelines/schematized-export-pipeline +/docs/other-bits/data-pipelines/schematized-export-pipeline /docs/data-pipelines/old-pipelines/schematized-export-pipeline /docs/other-bits/jql-overview /docs/reports/apps/jql /docs/other-bits/privacy-and-security /docs/privacy /docs/other-bits/privacy-and-security/eu-residency /docs/privacy/eu-residency @@ -237,7 +237,7 @@ /docs/tracking-methods/id-management/identifying-users /docs/tracking-methods/id-management/identifying-users-simplified /docs/data-pipelines/integrations/aws-raw-pipeline /docs/data-pipelines/integrations/raw-aws-pipeline /docs/data-pipelines/integrations/azure-raw-pipeline /docs/data-pipelines/integrations/raw-azure-pipeline -/docs/data-pipelines/integrations/amazon-s3 /docs/data-pipelines/integrations/schematized-aws-pipeline +/docs/data-pipelines/integrations/amazon-s3 /docs/data-pipelines/old-pipelines/integrations/schematized-aws-pipeline /docs/data-pipelines/integrations/google-cloud-storage-raw-pipeline /docs/data-pipelines/integrations/raw-gcs-pipeline /docs/data-pipelines/integrations/google-cloud-storage /docs/data-pipelines/integrations/schematized-gcs-pipeline /docs/features/session-replay /docs/session-replay