Data Lake Store
A no-limits data lake to power intelligent action
Unlock maximum value from all of your unstructured, semi-structured and structured data using the first cloud data lake built for enterprises with no limits on the size of data. Azure Data Lake Store is secured, massively scalable and built to the open HDFS standard, allowing you to run massively-parallel analytics.
Petabyte size files and Trillions of objects
With Azure Data Lake Store your organisation can analyse all of its data in a single place with no artificial constraints. Your Data Lake Store can store trillions of files where a single file can be greater than a petabyte in size which is 200x larger than other cloud stores. This makes Data Lake Store ideal for storing any type of data including massive datasets like high-resolution video, genomic and seismic datasets, medical data and data from a wide variety of industries.
Scalable throughput for massively parallel analytics
Without redesigning your application or repartitioning your data at higher scale, Data Lake Store scales throughput to support any size of analytic workload. It provides massive throughput to run analytic jobs with thousands of concurrent executors that read and write hundreds of terabytes of data efficiently.
Always encrypted, Role-based security and Auditing
Data Lake Store protects your data assets and extends your on-premises security and governance controls to the cloud easily. Data is always encrypted; in motion using SSL and at rest using service or user managed HSM-backed keys in Azure Key Vault. Capabilities such as single sign-on (SSO), multi-factor authentication and seamless management of millions of identities is built-in through Azure Active Directory. You can authorise users and groups with fine-grained POSIX-based ACLs for all data in the Store enabling role-based access controls. Finally, you can meet security and regulatory compliance needs by auditing every access or configuration change to the system.
HDFS for the Cloud
Microsoft Azure Data Lake Store supports any application that uses the open Apache Hadoop Distributed File System (HDFS) standard. By supporting HDFS, you can easily migrate your existing Hadoop and Spark data to the cloud without recreating your HDFS directory structure.
Apache Hadoop® and associated open source project names are trademarks of the Apache Software Foundation.
We guarantee a 99.9% enterprise-grade SLA and 24/7 support for your big data solution.