Persistent Identifier
|
doi:10.18710/4F4KJS |
Publication Date
|
2023-03-29 |
Title
| Supporting Data for: UltraMNIST Classification: A Benchmark to Train CNNs for Very Large Images |
Alternative URL
| https://www.kaggle.com/c/ultra-mnist |
Author
| Gupta, Deepak K. (UiT The Arctic University of Norway)
Bhamba, Udbhav
Thakur, Abhishek
Gupta, Akash
Sharan, Suraj
Demir, Ertugrul
Prasad, Dilip K. (UiT The Arctic University of Norway) - ORCID: 0000-0002-3693-6973 |
Point of Contact
|
Use email button above to contact.
Prasad, Dilip K. (UiT The Arctic University of Norway) |
Description
| Convolutional neural network (CNN) approaches available in the current literature are designed to work primarily with low-resolution images. When applied on very large images, challenges related to GPU memory, smaller receptive field than needed for semantic correspondence and the need to incorporate multi-scale features arise. The resolution of input images can be reduced, however, with significant loss of critical information. Based on the outlined issues, we introduce a novel research problem of training CNN models for very large images, and present ‘UltraMNIST dataset’, a simple yet representative benchmark dataset for this task. UltraMNIST has been designed using the popular MNIST digits with additional levels of complexity added to replicate well the challenges of real-world problems. We present two variants of the problem: ‘UltraMNIST classification’ and ‘Budget-aware UltraMNIST classification’. The standard UltraMNIST classification benchmark is intended to facilitate the development of novel CNN training methods that make the effective use of the best available GPU resources. The budget-aware variant is intended to promote development of methods that work under constrained GPU memory. For the development of competitive solutions, we present several baseline models for the standard benchmark and its budget-aware variant. We study the effect of reducing resolution on the performance and present results for baseline models involving pretrained backbones from among the popular state-of-the-art models. Finally, with the presented benchmark dataset and the baselines, we hope to pave the ground for a new generation of CNN methods suitable for handling large images in an efficient and resource-light manner. UltraMNIST dataset comprises very large-scale images, each of 4000x4000 pixels with 3-5 digits per image. Each of these digits has been extracted from the original MNIST dataset. Your task is to predict the sum of the digits per image, and this number can be anything from 0 to 27. (2022-04-15) |
Subject
| Medicine, Health and Life Sciences; Computer and Information Science |
Keyword
| Large Scale Image Classification
Ultra Large MNIST dataset
Variable scale features
CNN classification |
Related Publication
| Gupta, D. K., Bamba, U., Thakur, A., Gupta, A., Sharan, S., Demir, E., & Prasad, D. K. (2022). UltraMNIST Classification: A Benchmark to Train CNNs for Very Large Images. doi: 10.48550/arXiv.2206.12681 https://doi.org/10.48550/arXiv.2206.12681 |
Language
| English |
Producer
| UiT The Arctic University of Norway (UiT) https://en.uit.no/ |
Contributor
| Researcher : Gupta, Deepak K.
Researcher : Bhamba, Udbhav
Researcher : Thakur, Abhishek
Researcher : Gupta, Akash
Researcher : Sharan, Suraj
Researcher : Demir, Ertugrul
Researcher : Prasad, Dilip K.
Other : Nirwan Banerjee |
Funding Information
| The Research Council of Norway: grant no. 325741
UiT The Arctic University of Norway: Cristin project id 2061348 |
Distributor
| UiT The Arctic University of Norway (UiT The Arctic University of Norway) https://dataverse.no/dataverse/uit |
Depositor
| Banerjee, Nirwan |
Deposit Date
| 2023-03-06 |
Date of Collection
| Start Date: 2022-04-15 |
Data Type
| Image data; Large Scale data; Variable Scale Feature data |
Series
| Ultra-MNIST: https://www.kaggle.com/competitions/ultra-mnist |
Software
| Python, Version: 3
Numpy |
Data Source
| https://www.kaggle.com/competitions/ultra-mnist |