Read this in other languages: 简体中文
🚀 Deep Lake's efficient new C++ implementation speeds up data streaming by >2x compared to Hub 2.x (Ofeidis et al. 2022) 🚀
Deep Lake (formerly known as Activeloop Hub) is a data lake for deep learning applications. Open-source dataset format provides a simple API for creating, storing, and collaborating on AI datasets of any size. It enables you to store all of your data in one place, ranging from simple annotations to large videos, and it unlocks rapid streaming of data while training models at scale. Deep Lake is used by Google, Waymo, Red Cross, Omdena, Yale, & Oxford. Deep Lake includes the following features:
Storage Agnostic API
Use the same API to upload, download, and stream datasets to/from AWS S3/S3-compatible storage, GCP, Activeloop cloud, local storage, as well as in-memory.Compressed Storage
Store images, audios and videos in their native compression, decompressing them only when needed, for e.g., when training a model.Lazy NumPy-like Indexing
Treat your S3 or GCP datasets as if they are a collection of NumPy arrays in your system's memory. Slice them, index them, or iterate through them. Only the bytes you ask for will be downloaded!Dataset Version Control
Commits, branches, checkout - Concepts you are already familiar with in your code repositories can now be applied to your datasets as well!Integrations with Deep Learning Frameworks
Deep Lake comes with built-in integrations for Pytorch and Tensorflow. Train your model with a few lines of code - we even take care of dataset shuffling. :)Distributed Transformations
Rapidly apply transformations on your datasets using multi-threading, multi-processing, or our built-in Ray integration.100+ most-popular image, video, and audio datasets available in seconds
Deep Lake community has uploaded 100+ image, video and audio datasets like MNIST, COCO, ImageNet, CIFAR, GTZAN and others.Instant Visualization Support in Activeloop Platform
Deep Lake datasets are instantly visualized with bounding boxes, masks, annotations, etc. in Deep Lake Visualizer (see below).Deep Lake's core is efficiently built in C++ and can be quickly installed using pip.
pip3 install hub
By default, Deep Lake does not install dependencies for audio, video, and google-cloud (GCS) support. They can be installed using:
pip3 install "hub[av]" -> Audio and video support via PyAV
pip3 install "hub[gcp]" -> GCS support via google-* dependencies
pip3 install "hub[all]" -> Installs everything - audio, video and GCS support
import hub
import torch
from torchvision import transforms, models
ds = hub.load('hub://activeloop/cifar10-train')
ds.tensors.keys() # dict_keys(['images', 'labels'])
ds.labels[0].numpy() # array([6], dtype=uint32)
First, define a transform for the images and use Deep Lake's built-in PyTorch one-line dataloader to connect the data to the compute:
tform = transforms.Compose([
transforms.ToTensor(),
transforms.Normalize([0.5, 0.5, 0.5], [0.5, 0.5, 0.5]),
])
hub_loader = ds.pytorch(num_workers=0, batch_size=4, transform={
'images': tform, 'labels': None}, shuffle=True)
Next, define the model, loss and optimizer:
net = models.resnet18(pretrained=False)
net.fc = torch.nn.Linear(net.fc.in_features, len(ds.labels.info.class_names))
criterion = torch.nn.CrossEntropyLoss()
optimizer = torch.optim.SGD(net.parameters(), lr=0.001, momentum=0.9)
Finally, the training loop for 2 epochs:
for epoch in range(2):
running_loss = 0.0
for i, data in enumerate(hub_loader):
images, labels = data['images'], data['labels']
# zero the parameter gradients
optimizer.zero_grad()
# forward + backward + optimize
outputs = net(images)
loss = criterion(outputs, labels.reshape(-1))
loss.backward()
optimizer.step()
# print statistics
running_loss += loss.item()
if i % 100 == 99: # print every 100 mini-batches
print('[%d, %5d] loss: %.3f' %
(epoch + 1, i + 1, running_loss / 100))
running_loss = 0.0
A Deep Lake dataset can be created in various locations (Storage providers). This is how the paths for each of them would look like:
Storage provider | Example path |
---|---|
Activeloop cloud | hub://user_name/dataset_name |
AWS S3 / S3 compatible | s3://bucket_name/dataset_name |
GCP | gcp://bucket_name/dataset_name |
Google Drive | gdrive://path_to_dataset |
Local storage | path to local directory |
In-memory | mem://dataset_name |
Let's create a dataset in the Activeloop cloud. Activeloop cloud provides free storage up to 300 GB per user (more info here). Create a new account with Deep Lake from the terminal using activeloop register
if you haven't already. You will be asked for a user name, email ID, and password. The user name you enter here will be used in the dataset path.
$ activeloop register
Enter your details. Your password must be at least 6 characters long.
Username:
Email:
Password:
Initialize an empty dataset in the Activeloop Cloud:
import hub
ds = hub.empty("hub://<USERNAME>/test-dataset")
Next, create a tensor to hold images in the dataset we just initialized:
images = ds.create_tensor("images", htype="image", sample_compression="jpg")
Assuming you have a list of image file paths, let's upload them to the dataset:
image_paths = ...
with ds:
for image_path in image_paths:
image = hub.read(image_path)
ds.images.append(image)
Alternatively, you can also upload numpy arrays. Since the images
tensor was created with sample_compression="jpg"
, the arrays will be compressed with jpeg compression.
import numpy as np
with ds:
for _ in range(1000): # 1000 random images
random_image = np.random.randint(0, 256, (100, 100, 3)) # 100x100 image with 3 channels
ds.images.append(random_image)
You can load the dataset you just created with a single line of code:
import hub
ds = hub.load("hub://<USERNAME>/test-dataset")
You can also access one of the 100+ image, video and audio datasets in Deep Lake format, not just the ones you created. Here is how you would load the Objectron Bikes Dataset:
import hub
ds = hub.load('hub://activeloop/objectron_bike_train')
To get the first image in the Objectron Bikes dataset in numpy format:
image_arr = ds.image[0].numpy()
Getting started guides, examples, tutorials, API reference, and other useful information can be found on our documentation page.
Deep Lake users can access and visualize a variety of popular datasets through a free integration with Activeloop's Platform. Users can also create and store their own datasets and make them available to the public. Free storage of up to 300 GB is available for students and educators:
Storage for public datasets hosted by Activeloop | 200GB Free |
Storage for private datasets hosted by Activeloop | 100GB Free |
Deep Lake vs DVC
Deep Lake and DVC offer dataset version control similar to git for data, but their methods for storing data differ significantly. Deep Lake converts and stores data as chunked compressed arrays, which enables rapid streaming to ML models, whereas DVC operates on top of data stored in less efficient traditional file structures. The Deep Lake format makes dataset versioning significantly easier compared to traditional file structures by DVC when datasets are composed of many files (i.e., many images). An additional distinction is that DVC primarily uses a command-line interface, whereas Deep Lake is a Python package. Lastly, Deep Lake offers an API to easily connect datasets to ML frameworks and other common ML tools and enables instant dataset visualization through Activeloop's visualization tool.
Deep Lake vs TensorFlow Datasets (TFDS)
Deep Lake and TFDS seamlessly connect popular datasets to ML frameworks. Deep Lake datasets are compatible with both PyTorch and TensorFlow, whereas TFDS are only compatible with TensorFlow. A key difference between Deep Lake and TFDS is that Deep Lake datasets are designed for streaming from the cloud, whereas TFDS must be downloaded locally prior to use. As a result, with Deep Lake, one can import datasets directly from TensorFlow Datasets and stream them either to PyTorch or TensorFlow. In addition to providing access to popular publicly available datasets, Deep Lake also offers powerful tools for creating custom datasets, storing them on a variety of cloud storage providers, and collaborating with others via simple API. TFDS is primarily focused on giving the public easy access to commonly available datasets, and management of custom datasets is not the primary focus. A full comparison article can be found here.
Deep Lake vs HuggingFace
Deep Lake and HuggingFace offer access to popular datasets, but Deep Lake primarily focuses on computer vision, whereas HuggingFace focuses on natural language processing. HuggingFace Transforms and other computational tools for NLP are not analogous to features offered by Deep Lake.Deep Lake vs WebDatasets
Deep Lake and WebDatasets both offer rapid data streaming across networks. They have nearly identical steaming speeds because the underlying network requests and data structures are very similar. However, Deep Lake offers superior random access and shuffling, its simple API is in python instead of command-line, and Deep Lake enables simple indexing and modification of the dataset without having to recreate it.Deep Lake vs Zarr
Deep Lake and Zarr both offer storage of data as chunked arrays. However, Deep Lake is primarily designed for returning data as arrays using a simple API, rather than actually storing raw arrays (even though that's also possible). Deep Lake stores data in use-case-optimized formats, such as jpeg or png for images, or mp4 for video, and the user treats the data as if it's an array, because Deep Lake handles all the data processing in between. Deep Lake offers more flexibility for storing arrays with dynamic shape (ragged tensors), and it provides several features that are not naively available in Zarr such as version control, data streaming, and connecting data to ML Frameworks.Join our Slack community to learn more about unstructured dataset management using Deep Lake and to get help from the Activeloop team and other users.
We'd love your feedback by completing our 3-minute survey.
As always, thanks to our amazing contributors!
Made with contributors-img.
Please read CONTRIBUTING.md to get started with making contributions to Hub.
Using Deep Lake? Add a README badge to let everyone know:
[![deeplake](https://img.shields.io/badge/powered%20by-Deep%20Lake%20-ff5a1f.svg)](https://github.com/activeloopai/deeplake)
Dataset Licenses
Deep Lake users may have access to a variety of publicly available datasets. We do not host or distribute these datasets, vouch for their quality or fairness, or claim that you have a license to use the datasets. It is your responsibility to determine whether you have permission to use the datasets under their license.
If you're a dataset owner and do not want your dataset to be included in this library, please get in touch through a GitHub issue. Thank you for your contribution to the ML community!
Usage Tracking
By default, we collect usage data using Bugout (here's the code that does it). It does not collect user data other than anonymized IP address data, and it only logs the Deep Lake library's own actions. This helps our team understand how the tool is used and how to build features that matter to you! After you register with Activeloop, data is no longer anonymous. You can always opt-out of reporting using the CLI command below, or by setting an environmental variable BUGGER_OFF
to True
:
activeloop reporting --off
If you use Deep Lake in your research, please cite Activeloop using:
@article{deeplake,
doi = {10.48550/ARXIV.2209.10785},
url = {https://arxiv.org/abs/2209.10785},
author = {Hambardzumyan, Sasun and Tuli, Abhinav and Ghukasyan, Levon and Rahman, Fariz and Topchyan, Hrant and Isayan, David and Harutyunyan, Mikayel and Hakobyan, Tatevik and Stranic, Ivo and Buniatyan, Davit},
title = {Deep Lake: a Lakehouse for Deep Learning},
publisher = {arXiv},
year = {2022},
}
This technology was inspired by our research work at Princeton University. We would like to thank William Silversmith @SeungLab for his awesome cloud-volume tool.