.. _sec_automm_detection_prepare_pothole: AutoMM Detection - Prepare Pothole Dataset ========================================== `Pothole `__ is a small object detection dataset with 665 images, and has a specific domain, i.e. potholes on the road. This dataset will be used to show how to :ref:`sec_automm_detection_fast_ft_coco` and :ref:`sec_automm_detection_high_ft_coco`. You need 1 GB disk space to download and extract this dataset. SSD is preferred over HDD because of its better performance. The total time to prepare the dataset depends on your Internet speed and disk performance. For example, it often takes 3 min on AWS EC2 with EBS. You can download the dataset from its `kaggle page `__. Or you can also use our cli tool ``prepare_detection_dataset`` that can download all datasets mentioned in our tutorials. This python script is in our code: `prepare_detection_dataset.py `__, and you can also run it as a cli: ``python3 -m autogluon.multimodal.cli.prepare_detection_dataset``. Download with Python Script ~~~~~~~~~~~~~~~~~~~~~~~~~~~ You could either extract it in pothole folder under current directory by running: :: python3 -m autogluon.multimodal.cli.prepare_detection_dataset --dataset_name pothole or extract it in pothole folder under a provided output path: :: python3 -m autogluon.multimodal.cli.prepare_detection_dataset --dataset_name pothole --output_path ~/data or make it shorter: :: python3 -m autogluon.multimodal.cli.prepare_detection_dataset -d pothole -o ~/data The dataset downloaded with the provided tool is in COCO format and split to train/validation/test set with ratio 3:1:1. And the annotation files are: :: pothole/Annotations/usersplit_train_cocoformat.json pothole/Annotations/usersplit_val_cocoformat.json pothole/Annotations/usersplit_test_cocoformat.json .. _sec_automm_detection_prepare_coco17: and If You Download From Kaggle ~~~~~~~~~~~~~~~~~~~~~~~~~~~ Original Pothole dataset is in VOC format and is not split. **In Autogluon MultiModalPredictor, we strongly recommend using COCO as your data format instead. Check :ref:`convert_data_to_coco_format` for more information about COCO dataset and how to split and convert a VOC dataset to COCO.** Other Examples ~~~~~~~~~~~~~~ You may go to `AutoMM Examples `__ to explore other examples about AutoMM. Customization ~~~~~~~~~~~~~ To learn how to customize AutoMM, please refer to :ref:`sec_automm_customization`. Citation ~~~~~~~~ :: @inproceedings{inoue_2018_cvpr, author = {Inoue, Naoto and Furuta, Ryosuke and Yamasaki, Toshihiko and Aizawa, Kiyoharu}, title = {Cross-Domain Weakly-Supervised Object Detection Through Progressive Domain Adaptation}, booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, month = {June}, year = {2018} }