Nowadays, researchers are striving to implement human-level understanding into robots. For instance, understanding the surrounding environment, hand movement, detecting and grasping objects and many more.
In this article, we list down 10 open-source datasets that can be used to learn robotics in an advanced way.
1| Robot Arm Pushing Dataset
The robot arm pushing dataset contains approximately 59,000 examples of robot pushing motions which including one training set (train) and two test sets. The first test set used two different subsets of the objects pushed during training. The second test set involved two subsets of objects. One key application of this dataset is to implement it in a learned model for decision making in vision-based robotic control tasks.
2| Robot Arm Grasping Dataset
The robot arm grasping dataset contains approximately 650,000 examples of robot grasping attempts. The dataset was mainly created to train a large convolutional neural network (CNN) to predict the probability that the task-space motion of the gripper will result in successful grasps.
3| Dataset Of Daily Interactive Manipulation
The dataset of daily interactive manipulation focuses on position, orientation, force, and torque of objects manipulated in daily tasks. It is a collection of 3D position and orientation (PO), force and torque (FT) data of tools/objects being manipulated to fulfill certain tasks. It includes 1,603 trials of 32 types of daily motions and 1,596 trials of pouring alone along with the helper code.
4| Fukuoka Datasets For Place Categorisation
Fukuoka datasets for place categorisation is a collection of several multimodal 3D datasets. It contains several datasets for the task of place categorisation using global information without object segmentation. It can also be used for other segmentation-based methods and can be extended with additional annotations. The datasets include indoor and outdoor scenarios obtained in different locations in Fukuoka city, Japan.
5| MultiDrone Public DataSet
Aerial robots can be used for many robotic tasks, for instance, surveillance, monitoring, filming, among others. The MultiDrone public dataset has been collected using both pre-existing audiovisual material and newly filmed UAV shots. It includes more than 10 hours of UAV footage depicting football, rowing, and cycling (DW). A large subset of this dataset has been annotated for facilitating scientific research, in tasks such as visual detection and tracking of bicycles, football players, human crowds, etc.
6| Complex Urban Dataset With Multi-Level Sensors From Highly Diverse Urban Environments
The complex urban dataset with multi-level sensors from a highly diverse urban environment captures various complex urban features and addresses the major issues of complex urban areas, such as unreliable and sporadic Global Positioning System (GPS) data, multi-lane roads, complex building structures, and the abundance of highly dynamic objects. In this dataset, typical LiDAR sensor data, 2D and 3D LiDAR, are provided.
7| Natural Language Instructions For Human-Robot Collaborative Manipulation
Natural language instructions for human-robot collaborative manipulation is a dataset of natural language instructions for object reference in manipulation scenarios. It comprises of 1582 individual written instructions which were collected via online crowdsourcing. This dataset is particularly useful for researchers who work in natural language processing, human-robot interaction, and robotic manipulation.
8| Yale-CMU-Berkeley Dataset For Robotic Manipulation Research
Yale-CMU-Berkeley dataset for robotic manipulation research is an image and model dataset of the real-life objects from the Yale-CMU-Berkeley Object Set, which is specifically designed for benchmarking in manipulation research. The dataset includes 600 high-resolution RGB images, 600 RGB-D images and five sets of textured three-dimensional geometric models.
9| Robot-at-Home Dataset
Robot-at-home is a robotic dataset for semantic mapping of home environments. It is a collection of raw and processed sensory data from domestic settings which are used for semantic mapping algorithms through the categorisation of objects or rooms. The dataset contains 87,000+ time-stamped observations gathered by a mobile robot endowed with a rig of four RGB-D cameras and a 2D laser scanner.
10| Robotic 3D Scan Repository
Robotic 3D scan repository is a data repository of standard 3D data sets for the robotics community, log files of robot runs and 3D point clouds from robotic experiments. One can easily use this data for developing SLAM or interpretation algorithms.