Data Lake Store

A no-limits data lake to power intelligent action

Unlock the most value from all of your unstructured, semi-structured and structured data using the first cloud data lake built for enterprises, with no limits on the size of data. Azure Data Lake Store is secured, massively scalable and built in accordance with the open HDFS standard, allowing you to run massively parallel analytics.

Petabyte-size files and trillions of objects

With Azure Data Lake Store, your organisation can analyse all of its data in one place, with no artificial constraints. Your Data Lake Store can store trillions of files, and a single file can be greater than a petabyte in size – 200 times larger than other cloud stores. This makes Data Lake Store ideal for storing any type of data, including massive data sets such as high-resolution video, genomic and seismic data sets, medical data and data from a wide variety of industries.

Scalable throughput for massively parallel analytics

Without redesigning your application or repartitioning your data at a higher scale, Data Lake Store scales throughput to support any size of analytic workload. It provides massive throughput to run analytic jobs, with thousands of concurrent executors that read and write hundreds of terabytes of data efficiently.

Always-encrypted, role-based security & auditing

Data Lake Store protects your data assets and extends your on-premises security and governance controls to the cloud easily. Data is always encrypted – in motion using SSL, and at rest using service or user-managed HSM-backed keys in Azure Key Vault. Capabilities such as single sign-on (SSO), multi-factor authentication and seamless management of millions of identities are built in with Azure Active Directory. You can authorise users and groups with fine-grained POSIX-based ACLs for all data in the Store, enabling role-based access controls. Finally, you can meet security and regulatory compliance needs by auditing every access or configuration change to the system.

HDFS for the Cloud

Microsoft Azure Data Lake Store supports any application that uses the open Apache Hadoop Distributed File System (HDFS) standard. By supporting HDFS, you can easily migrate your existing Hadoop and Spark data to the cloud without recreating your HDFS directory structure.

Apache Hadoop® and associated open-source project names are trademarks of the Apache Software Foundation.

Enterprise-grade Support

We guarantee a 99.9% enterprise-grade SLA and 24/7 support for your big data solution.

Related products and services

Data Lake Analytics

Distributed analytics service that makes big data easy


Provision cloud Hadoop, Spark, R Server, HBase and Storm clusters

Try Data Lake Store