google-research-datasets / Objectron

Licence: other
Objectron is a dataset of short, object-centric video clips. In addition, the videos also contain AR session metadata including camera poses, sparse point-clouds and planes. In each video, the camera moves around and above the object and captures it from different views. Each object is annotated with a 3D bounding box. The 3D bounding box describes the object’s position, orientation, and dimensions. The dataset contains about 15K annotated video clips and 4M annotated images in the following categories: bikes, books, bottles, cameras, cereal boxes, chairs, cups, laptops, and shoes

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Objectron

Cubicasa5k
CubiCasa5k floor plan dataset
Stars: ✭ 98 (-92.75%)
Mutual labels:  jupyter-notebook, dataset
Pymeshfix
Python Wrapper for MeshFix: easily repair holes in PyVista surface meshes
Stars: ✭ 75 (-94.45%)
Mutual labels:  3d, 3d-reconstruction
Vitech
tuyển chọn các tài liệu về công nghệ bằng tiếng Việt
Stars: ✭ 63 (-95.34%)
Mutual labels:  ai, jupyter-notebook
Animegan
A simple PyTorch Implementation of Generative Adversarial Networks, focusing on anime face drawing.
Stars: ✭ 1,095 (-19.01%)
Mutual labels:  jupyter-notebook, dataset
Eval Vislam
Toolkit for VI-SLAM evaluation.
Stars: ✭ 89 (-93.42%)
Mutual labels:  dataset, augmented-reality
Wikipedia ner
📖 Labeled examples from wiki dumps in Python
Stars: ✭ 61 (-95.49%)
Mutual labels:  jupyter-notebook, dataset
Raccoon dataset
The dataset is used to train my own raccoon detector and I blogged about it on Medium
Stars: ✭ 1,177 (-12.94%)
Mutual labels:  jupyter-notebook, dataset
Teacher Student Training
This repository stores the files used for my summer internship's work on "teacher-student learning", an experimental method for training deep neural networks using a trained teacher model.
Stars: ✭ 34 (-97.49%)
Mutual labels:  ai, jupyter-notebook
Openml R
R package to interface with OpenML
Stars: ✭ 81 (-94.01%)
Mutual labels:  jupyter-notebook, dataset
Pragmaticai
[Book-2019] Pragmatic AI: An Introduction to Cloud-based Machine Learning
Stars: ✭ 79 (-94.16%)
Mutual labels:  ai, jupyter-notebook
Cinemanet
Stars: ✭ 57 (-95.78%)
Mutual labels:  jupyter-notebook, dataset
Ai Dl Enthusiasts Meetup
AI & Deep Learning Enthusiasts Meetup Project & Study Sessions
Stars: ✭ 90 (-93.34%)
Mutual labels:  ai, jupyter-notebook
Covidnet Ct
COVID-Net Open Source Initiative - Models and Data for COVID-19 Detection in Chest CT
Stars: ✭ 57 (-95.78%)
Mutual labels:  jupyter-notebook, dataset
Aiopen
AIOpen是一个按人工智能三要素(数据、算法、算力)进行AI开源项目分类的汇集项目,项目致力于跟踪目前人工智能(AI)的深度学习(DL)开源项目,并尽可能地罗列目前的开源项目,同时加入了一些曾经研究过的代码。通过这些开源项目,使初次接触AI的人们对人工智能(深度学习)有更清晰和更全面的了解。
Stars: ✭ 62 (-95.41%)
Mutual labels:  ai, jupyter-notebook
Chinesetrafficpolicepose
Detects Chinese traffic police commanding poses 检测中国交警指挥手势
Stars: ✭ 49 (-96.38%)
Mutual labels:  jupyter-notebook, dataset
Toronto 3d
A Large-scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways
Stars: ✭ 69 (-94.9%)
Mutual labels:  dataset, 3d
Tedsds
Apache Spark - Turbofan Engine Degradation Simulation Data Set example in Apache Spark
Stars: ✭ 14 (-98.96%)
Mutual labels:  jupyter-notebook, dataset
Deep learning projects
Stars: ✭ 28 (-97.93%)
Mutual labels:  jupyter-notebook, dataset
Symbolic Musical Datasets
🎹 symbolic musical datasets
Stars: ✭ 79 (-94.16%)
Mutual labels:  jupyter-notebook, dataset
Deep Dream In Pytorch
Pytorch implementation of the DeepDream computer vision algorithm
Stars: ✭ 90 (-93.34%)
Mutual labels:  ai, jupyter-notebook

Objectron Dataset

Objectron is a dataset of short object centric video clips with pose annotations.


WebsiteDataset FormatTutorialsLicense

The Objectron dataset is a collection of short, object-centric video clips, which are accompanied by AR session metadata that includes camera poses, sparse point-clouds and characterization of the planar surfaces in the surrounding environment. In each video, the camera moves around the object, capturing it from different angles. The data also contain manually annotated 3D bounding boxes for each object, which describe the object’s position, orientation, and dimensions. The dataset consists of 15K annotated video clips supplemented with over 4M annotated images in the following categories: bikes, books, bottles, cameras, cereal boxes, chairs, cups, laptops, and shoes. In addition, to ensure geo-diversity, our dataset is collected from 10 countries across five continents. Along with the dataset, we are also sharing a 3D object detection solution for four categories of objects — shoes, chairs, mugs, and cameras. These models are trained using this dataset, and are released in MediaPipe, Google's open source framework for cross-platform customizable ML solutions for live and streaming media.

Key Features

  • 15000 annotated videos and 4M annotated images
  • All samples include high-res images, object pose, camera pose, point-cloud, and surface planes.
  • Ready to use examples in various tf.record formats, which can be used in Tensorflow/PyTorch.
  • Object-centric multi-views, observing the same object from different angles.
  • Accurate evaluation metrics, like 3D IoU for oriented 3D bounding boxes.

Dataset Format

The data is stored in the objectron bucket on Google Cloud storage. Check out the Download Data notebook for a quick review of how to download/access the dataset. The following assets are available:

  • The video sequences (located in /videos/class/batch-i/j/video.MOV files)
  • The annotation labels containing the 3D bounding boxes for objects. The annotation protobufs are located in /videos/class/batch-i/j/geometry.pbdata files. They are formatted using the object.proto. See [example] on how to parse the annotation files.
  • AR metadata (such as camera poses, point clouds, and planar surfaces). They are based on a_r_capture_metadata.proto. See example on how to parse these files.
  • Processed dataset: sharded and shuffled tf.records of the annotated frames, in tf.example format and videos in tf.SequenceExample format. These are used for creating the input data pipeline to your models. These files are located in /v1/records_shuffled/class/ and /v1/sequences/class/.
  • Supporting scripts to run evaluation based on the 3D IoU metric.
  • Supporting scripts to load the data into Tensorflow, Jax and Pytorch and visualize the dataset, including “Hello World” examples.
  • Supporting Apache Beam jobs to process the datasets on Google Cloud infrastructure.
  • The index of all available samples, as well as train/test splits for easy access and download.

Raw dataset size is 1.9TB (including videos and their annotations). Total dataset size is 4.4TB (including videos, records, sequences, etc.). This repository provides the required schemas and tools to parse the dataset.

class bike book bottle camera cereal_box chair cup laptop shoe
#videos 476 2024 1928 815 1609 1943 2204 1473 2116
#frames 150k 576k 476k 233k 396k 488k 546k 485k 557k

Tutorials

License

Objectron is released under Computational Use of Data Agreement 1.0 (C-UDA-1.0). A copy of the license is available in this repository.

BibTeX

If you found this dataset useful, please cite our paper.

@article{objectron2020,
  title={Objectron: A Large Scale Dataset of Object-Centric Videos in the Wild with Pose Annotations},
  author={Adel Ahmadyan, Liangkai Zhang, Jianing Wei, Artsiom Ablavatski, Matthias Grundmann},
  journal={arXiv preprint arXiv:2012.09988},
  year={2020}
}

This is not an officially supported Google product. If you have any question, you can email us at [email protected] or join our mailing list at [email protected]

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].