Solutions
Datasets
Download CSV, ORC, and Parquet data files.
Analytics
Connect your BI tools to our analytical query service.
Integrations
Enhance your analytics solutions with our datasets.
Insights
Interactive reports with actionable insights.
Use Cases
Learn how you could unlock value from our datasets.
Consulting
Transform our datasets into your competitive edge.
PricingAboutContact
Resources
Help Centre
Find answers to the most frequently asked questions.
Documentation
Learn everything you need to know about Open Data Blend.
Blog
Keep up to date with our latest news, updates, and thoughts.
Get Involved
Help to improve the Open Data Blend services for everyone.
Affiliates
Supplement your business with a new recurring revenue stream.
Manage Subscription
Solutions
Datasets
Download CSV, ORC, and Parquet data files.
Analytics
Connect your BI tools to our analytical query service.
Integrations
Enhance your analytics solutions with our datasets.
Insights
Interactive reports with actionable insights.
Use Cases
Learn how you could unlock value from our datasets.
Consulting
Transform our datasets into your competitive edge.
PricingAboutContact
Resources
Help Center
Find answers to the most frequently asked questions.
Documentation
Learn everything you need to know about Open Data Blend.
Blog
Keep up to date with our latest news, updates, and thoughts.
Get Involved
Help to improve the Open Data Blend services for everyone.
Affiliates
Supplement your business with a new recurring revenue stream.
Manage Subscription

Open Data Blend February 2022 Update

Recent articles
Open Data Blend February 2023 Update
10th March 2023
Open Data Blend January 2023 Update
10th February 2023
6 Traits of a Great Data Engineer
26th January 2023
8 Traits of a Great Data Scientist
12th January 2023
Open Data Blend December 2022 Update
3rd January 2023

18th March 2022

By Open Data Blend Team

The Open Data Blend February 2022 update includes changes that extend the Open Data Blend for Python library to support cloud data lake storage services.

Open Data Blend Datasets


English Prescribing Data for January 2022 Is Available

We have updated the Prescribing dataset with the latest available NHS English Prescribing Data which includes activity up until January 2022. You can download the data from the Open Data Blend Datasets Prescribing page, analyse it directly in supported BI tools through the Open Data Blend Analytics service, or instantly explore insights through the Open Data Blend Insights service.

Cloud Data Lake Integrations with Open Data Blend for Python


We have extended Open Data Blend for Python to support Azure Blob Storage, Azure Data Lake Storage (ADLS) Gen2, and Amazon S3 as target file systems.

Open Data Blend for Python Data Lake Support

With a few simple lines of code, you can quickly ingest our datasets into your data lake. Once ingested, you can interactively query and analyse the ORC and Parquet files using data lake analytics services like Amazon Athena, Azure Synapse Analytics, and Databricks.

Below you can see some examples of how to use the new cloud data lake integrations.

Ingesting Data Directly into Azure Blob Storage

import opendatablend as odb

dataset_path = 'https://packages.opendatablend.io/v1/open-data-blend-road-safety/datapackage.json'
access_key = '<ACCESS_KEY>' # The access key can be set to an empty string if you are making a public API request

# Specify the resource name of the data file. In this example, the 'date' data file will be requested in .parquet format.
resource_name = 'date-parquet'

# Get the data and store the output object using the Azure Blob Storage file system
configuration = {
    "connection_string" : "DefaultEndpointsProtocol=https;AccountName=<AZURE_BLOB_STORAGE_ACCOUNT_NAME>;AccountKey=<AZURE_BLOB_STORAGE_ACCOUNT_KEY>;EndpointSuffix=core.windows.net",
    "container_name" : "<AZURE_BLOB_STORAGE_CONTAINER_NAME>" # e.g. odbp-integration
}
output = odb.get_data(dataset_path, resource_name, access_key=access_key, file_system="azure_blob_storage", configuration=configuration)

# Print the file locations
print(output.data_file_name)
print(output.metadata_file_name)

Ingesting Data Directly into Azure Data Lake Storage (ADLS) Gen2

import opendatablend as odb

dataset_path = 'https://packages.opendatablend.io/v1/open-data-blend-road-safety/datapackage.json'
access_key = '<ACCESS_KEY>' # The access key can be set to an empty string if you are making a public API request

# Specify the resource name of the data file. In this example, the 'date' data file will be requested in .parquet format.
resource_name = 'date-parquet'

# Get the data and store the output object using the Azure Data Lake Storage Gen2 file system
configuration = {
    "connection_string" : "DefaultEndpointsProtocol=https;AccountName=<ADLS_GEN2_ACCOUNT_NAME>;AccountKey=<ADLS_GEN2_ACCOUNT_KEY>;EndpointSuffix=core.windows.net",
    "container_name" : "<ADLS_GEN2_CONTAINER_NAME>" # e.g. odbp-integration
}
output = odb.get_data(dataset_path, resource_name, access_key=access_key, file_system="azure_blob_storage", configuration=configuration)

# Print the file locations
print(output.data_file_name)
print(output.metadata_file_name)

Ingesting Data Directly into Amazon S3

import opendatablend as odb

dataset_path = 'https://packages.opendatablend.io/v1/open-data-blend-road-safety/datapackage.json'
access_key = '<ACCESS_KEY>' # The access key can be set to an empty string if you are making a public API request

# Specify the resource name of the data file. In this example, the 'date' data file will be requested in .parquet format.
resource_name = 'date-parquet'

# Get the data and store the output object using the Amazon S3 file system
configuration = {
    "aws_access_key_id" : "<AWS_ACCESS_KEY_ID>",
    "aws_secret_access_key" : "AWS_SECRET_ACCESS_KEY",
    "bucket_name" : "<BUCKET_NAME>", # e.g. odbp-integration
    "bucket_region" : "<BUCKET_REGION>" # e.g. eu-west-2
}

output = odb.get_data(dataset_path, resource_name, access_key=access_key, file_system="amazon_s3", configuration=configuration)

# Print the file locations
print(output.data_file_name)
print(output.metadata_file_name)

Want to learn more about how Open Data Blend for Python can help you to integrate our datasets? Head over to the GitHub or PyPI page.

Follow Us and Stay Up to Date

Keep up to date with Open Data Blend by following us on Twitter and LinkedIn. Be among the first to know when there's something new.

Blog hero image by Natasha Miller on Unsplash.

Got feedback?
Get involved.
Get our latest updates
We'll use the information you provide through this form to send you Open Data Blend related news and updates. View our privacy policy
Operated by

Copyright © 2019-2023 Nimble Learn Ltd. All rights reserved unless otherwise stated. Company Registration Number 08637310. VAT Number 174 9728 60.

Open Data Blend®, the Open Data Blend® logo, Nimble Learn®, and the Nimble Learn® logo are registered trademarks of Nimble Learn Ltd. All other product names, logos, and brands are the property of their respective owners, and their use does not imply endorsement.

Terms    Privacy    Cookies    SLA    Licensing    Docs    Status