To populate your GCP + Google Cloud Storage (GCS) repository with healthcare data from an EHR system via Redox (and then to optionally feed that data into Databricks or Snowflake for analytics), you must configure a specific Redox cloud destination. A Redox destination represents where a message is delivered (e.g., like the address in the “To” line of an email header). Learn more about connecting Redox to your cloud repository.
You'll need to perform some steps in your cloud product(s) and some in Redox. You can perform Redox setup in our dashboard or with the Redox Platform API.
- Establish a connection with your preferred EHR system. Learn how to request a connection.
- Decide which combination of cloud products to use. Redox currently supports any of these combinations with your GCP cloud repository:
- Google Cloud Platform (GCP) + Google Cloud Storage (GCS)
- GCP + GCS + Databricks
- GCP + GCS + Snowflake
- Complete your GCP (and any other cloud product) configuration before creating your Redox destination. Save any downloads with secret values, since you’ll need to enter some of these details into the Redox dashboard.
- Grant access to Redox from GCP (and any other cloud product) to authorize Redox to push data to your cloud repository.
- Navigate to the GCP dashboard and log in.
- Create a service account in your GCP project.
- Grant the service account read/write access to the GCS bucket.
- Create a new key for your service account.
- Download the new key file. Save this file, since you’ll need it for Redox setup later.
Next, create a cloud destination in your Redox organization. When the EHR system sends healthcare data to Redox, we push it on to your configured GCP + GCS cloud destination.
- For the select destination step, follow the instructions for creating a cloud destination.
- From the Product type field, select Databricks or Snowflake if you’re using one of those cloud products with GCS. Your GCS settings will be applied with the additional cloud product. Select Cloud Storage if you’re not using either Databricks or Snowflake.
- For the configure destination step, populate these fields.
- Bucket name: Enter the GCS bucket name. Locate this value in the GCP dashboard.Bucket naming requirements
- Object filename prefix: Enter any prefix you want prepended to new files when they’re created in the GCS bucket. Add / to put the files in a subdirectory. For example, redox/ puts all the files in the redox directory.
- Click the Next button.
- For the auth credential step, either a drop-down list of existing auth credentials displays or a new auth credential form opens. Learn how to create an auth credential for OAuth 2.0 2-legged with JWT for GCP.Existing or new auth credentialDifferent Client ID
- For the verify step, follow the instructions for verifying a destination.
- In your terminal, prepare the /v1/authcredentials request.
- Specify these values in the request.
- Locate the private_key_id, private_key_id, and client_email values in the downloaded key file from GCP.
- Make sure the auth credential’s environment ID matches the environment ID of the destination you intend to link it to.
- Use the values we provide in the example for authStrategy, grantType, algorithm, url, and scopes. These are fixed values.Example: Create auth credential for GCP + GCS + Databricks or Snowflakejson
- You should get a successful 200 response and a payload populated with the details of the new auth credential.
- In your terminal, prepare the /v1/environments/{environmentId}/destinations request with these values:
- Set authCredential to the auth credential ID from the response you received in step #4.
- Populate cloudProviderSettings with the settings below.
- The fileNamePrefix is optional. If specified, the filename format will be the prefix you define appended by the log ID. If not specified, the filename will simply be the log ID.
- Enter the productId based on your specific setup:
- GCS only: gcs.
- GCS + Databricks: databricks
- GCS + Snowflake: snowflakeExample: Values for GCP + Databricks or Snowflake cloudProviderSettingsjson
- You should get a successful 200 response with a payload populated with the details of the new GCP cloud destination. Specifically, the verified status of the destination should be set to true.
- Your new destination will now be able to receive messages. These messages get stored in your GCS bucket, which can then be picked up by Databricks or Snowflake for further processing.