Data Lake Store
A no-limits data lake to power intelligent action
Unlock the most value from all of your unstructured, semi-structured and structured data using the first cloud data lake built for enterprises, with no limits on the size of data. Azure Data Lake Store is secured, massively scalable and built in accordance with the open HDFS standard, allowing you to run massively parallel analytics.
Petabyte-size files and trillions of objects
With Azure Data Lake Store, analyse all your data in one place with no artificial constraints. Data Lake Store can store trillions of files. A single file can be larger than one petabyte in size – 200 times larger than other cloud store options. This makes Data Lake Store ideal for storing any type of data, including massive data sets such as high-resolution video, genomic and seismic data sets, medical data and data from a wide variety of industries.
Scalable throughput for massively parallel analytics
Without you needing to redesign your application or repartition your data at higher scale, Data Lake Store scales throughput to support any size of analytic workload. You get massive throughput to run analytic jobs with thousands of concurrent executors that efficiently read and write hundreds of terabytes of data.
Always encrypted, role-based security and auditing
Data Lake Store protects your data assets and extends your on-premises security and governance controls to the cloud. Your data is always encrypted, while in motion using SSL, and at rest using service or user-managed HSM-backed keys in Azure Key Vault. Single sign-on (SSO), multi-factor authentication and seamless management of millions of identities is built in through Azure Active Directory. Authorise users and groups with fine-grained POSIX-based ACLs for all data in your store and enable role-based access controls. Meet security and regulatory compliance needs by auditing every access or configuration change to the system.
HDFS for the cloud
Microsoft Azure Data Lake Store supports any application that uses the open Apache Hadoop Distributed File System (HDFS) standard. By supporting HDFS, you can easily migrate your existing Hadoop and Spark data to the cloud without recreating your HDFS directory structure.
Apache Hadoop® and associated open-source project names are trademarks of the Apache Software Foundation.
With Data Lake Store, you get a 99.9% enterprise-grade SLA and 24/7 support for your big data solution.