The Azure Data Lake Integration provides a way to import your data from one of the most widely used cloud file storages into the Bloomreach Engagement to have your customer and event data in one place and ready to analyze them in the app fully.
Azure Storage integration also supports Azure Data Lake Gen2 type of storage accounts. Please refer to Azure documentation for more details on the Data Lake Gen2 type of storage accounts.
Exports to Azure Storage
The integration handles imports from Azure Storage. If you wish to export your data from Bloomreach Engagement to Azure Storage, use the separate Azure Storage Integration for Exports.
The Azure Storage Integration for Imports can be added in
Data & Assets ->
Add new integration. After that, you need to enter your Azure Storage account name, container name, Application (client) ID, Directory (tenant) ID, and your client secret. You need to get all of this information from your Azure Portal account.
If you want to use the integration, you need:
- Azure storage account
- Register a new application in your Azure account
- Configure read permissions for your app for your storage account
- Sign in to the Azure Portal.
- Decide which container in which storage account should be used as a data source for your imports. Then copy the storage account name and container name into the respective fields.
- Register the application in the
App registrationsection of the Azure Portal.
- Copy the Application (client) ID and Directory (tenant) ID to the respective fields.
- Create a new client secret in the
Certifications & Secretssection of your app registration.
- Then copy your newly created client secret into the respective field.
Containersand choose the container you want to use as a source for your import.
- Copy the container name from the respective field.
- In the
Access Control (IAM)section of your storage account, assign the Storage Blob Data Reader role to the app registration you created in the third step.
To register the application in the App registration section, you need to have assigned roles with permission for the Azure Blob Storage Reader.
- Authorization is currently only possible using OAuth2.
Keep in mind that your client's secret has an expiration date set in the future. Once the token expires, your import may stop working due to authentication issues.
I keep getting the error that the endpoint does not support BlobStorageEvents or SoftDelete.
- Go to the settings of your Azure storage and turn off the soft-delete feature.
Updated about 1 month ago