< Previous Challenge - Home - Next Challenge>
In this challenge, you will explore bulk exporting, anonymizing and storing FHIR data into Data Lake.
The FHIR Tools for Anonymization is an open-source project that helps anonymize healthcare FHIR data, on-premises or in the cloud, for secondary usage such as research, public health, and more. This architecture uses multiple Azure services for creating an automated pipeline to process the bulk export and anonymization for FHIR. The goal of the template is to enable quick and continuous creation of research datasets while applying HIPAA safe harbor rules.
A Timer is used to trigger the Logic App which bulk exports data from FHIR and stores in a preset storage location. The Logic App loops on an adjustable 5 minute interval until Bulk Export finishes exporting all data from FHIR. Logic App runs Azure Data Factory which in triggers Azure Batch which performs the deidentification with the FHIR Tools for Anonymization. The deidentified data is stored in Azure Data Lake Gen 2 for further use.
You will deploy using the Microsoft Health Architectures.
health-architectures/Research-and-Analytics/FHIRExportwithAnonymization
and open the ./Assets/arm_template_parameters.json
file in your preferred JSON editor. Replace FHIR URL, client id, client secret, tenant id and export storage account with yours. Connect-AzAccount
Get-AzSubscription
Select-AzSubscription -SubscriptionId "<SubscriptionId>"
cd health-architectures-master\Research-and-Analytics\FHIRExportwithAnonymization
$EnvironmentName = "<NAME HERE>" #The name must be lowercase, begin with a letter, end with a letter or digit, and not contain hyphens.
$EnvironmentLocation = "<LOCATION HERE>" #optional input. The default is eastus2
./deployFHIRExportwithAnonymization.ps1 -EnvironmentName $EnvironmentName -EnvironmentLocation $EnvironmentLocation #Environment Location is optional
{ENVIRONMENTNAME}
{ENVIRONMENTNAME}adf
{ENVIRONMENTNAME}batch
{ENVIRONMENTNAME}kv
{ENVIRONMENTNAME}la
{ENVIRONMENTNAME}dlg2
{ENVIRONMENTNAME}kv
deployed with the script. Open the key vault, locate ‘Secrets’ blade under ‘Settings’. Click on the secret named ‘blobstorageacctstring’. Then click “+ New Version”. In the ‘Value’ box paste the connection string from the storage account. Then click the ‘Create’ button at the bottom the page. This will point the Azure Data Factory to the pre-configured FHIR Integration Storage Account.{ENVIRONMENTNAME}la
deployed with the script and click Run Trigger. Click on the Running status in Runs History below in the same screen. The time taken to complete depends on the volume of data you have in Azure API for FHIR.{ENVIRONMENTNAME}dlg2
deployed with the script. Look for the container with output as suffix.