SyNet: an ensemble network for object detection in UAV images

buir.contributor.authorAlbaba, Berat Mert
buir.contributor.authorÖzer, Sedat
dc.citation.epage10234en_US
dc.citation.spage10227en_US
dc.contributor.authorAlbaba, Berat Mert
dc.contributor.authorÖzer, Sedat
dc.coverage.spatialMilan, Italyen_US
dc.date.accessioned2022-02-08T10:08:11Z
dc.date.available2022-02-08T10:08:11Z
dc.date.issued2021-05-05
dc.departmentDepartment of Computer Engineeringen_US
dc.departmentDepartment of Mechanical Engineeringen_US
dc.descriptionConference Name: 2020 25th International Conference on Pattern Recognition (ICPR)en_US
dc.descriptionDate of Conference: 10-15 Jan. 2021en_US
dc.description.abstractRecent advances in camera equipped drone applications and their widespread use increased the demand on vision based object detection algorithms for aerial images. Object detection process is inherently a challenging task as a generic computer vision problem, however, since the use of object detection algorithms on UAVs (or on drones) is relatively a new area, it remains as a more challenging problem to detect objects in aerial images. There are several reasons for that including: (i) the lack of large drone datasets including large object variance, (ii) the large orientation and scale variance in drone images when compared to the ground images, and (iii) the difference in texture and shape features between the ground and the aerial images. Deep learning based object detection algorithms can be classified under two main categories: (a) single-stage detectors and (b) multi-stage detectors. Both single-stage and multi-stage solutions have their advantages and disadvantages over each other. However, a technique to combine the good sides of each of those solutions could yield even a stronger solution than each of those solutions individually. In this paper, we propose an ensemble network, SyNet, that combines a multi-stage method with a single-stage one with the motivation of decreasing the high false negative rate of multi-stage detectors and increasing the quality of the single-stage detector proposals. As building blocks, CenterNet and Cascade R-CNN with pretrained feature extractors are utilized along with an ensembling strategy. We report the state of the art results obtained by our proposed solution on two different datasets: namely MS-COCO and visDrone with %52.1 mAP IoU=0.75 is obtained on MS-COCO val2017 dataset and %26.2 mAP IoU=0.75 is obtained on VisDrone test - set. Our code is available at: https://github.com/mertalbaba/SyNet.en_US
dc.description.provenanceSubmitted by Betül Özen (ozen@bilkent.edu.tr) on 2022-02-08T10:08:11Z No. of bitstreams: 1 SyNet_An_Ensemble_Network_for_Object_Detection_in_UAV_Images.pdf: 4957786 bytes, checksum: 6b5e563cfaca7328edac54a62b089e78 (MD5)en
dc.description.provenanceMade available in DSpace on 2022-02-08T10:08:11Z (GMT). No. of bitstreams: 1 SyNet_An_Ensemble_Network_for_Object_Detection_in_UAV_Images.pdf: 4957786 bytes, checksum: 6b5e563cfaca7328edac54a62b089e78 (MD5) Previous issue date: 2021-05-05en
dc.identifier.doi10.1109/ICPR48806.2021.9412847en_US
dc.identifier.eisbn978-1-7281-8808-9en_US
dc.identifier.isbn978-1-7281-8809-6en_US
dc.identifier.issn1051-4651en_US
dc.identifier.urihttp://hdl.handle.net/11693/77130en_US
dc.language.isoEnglishen_US
dc.publisherIEEEen_US
dc.relation.isversionofhttps://dx.doi.org/10.1109/ICPR48806.2021.9412847en_US
dc.source.titleInternational Conference on Pattern Recognitionen_US
dc.subjectDeep learningen_US
dc.subjectEnsemble methodsen_US
dc.subjectObject detectionen_US
dc.subjectUAV imagesen_US
dc.titleSyNet: an ensemble network for object detection in UAV imagesen_US
dc.typeConference Paperen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
SyNet_An_Ensemble_Network_for_Object_Detection_in_UAV_Images.pdf
Size:
4.73 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.69 KB
Format:
Item-specific license agreed upon to submission
Description: