OU Portal
Log In
Welcome
Applicants
Z6_60GI02O0O8IDC0QEJUJ26TJDI4
Error:
Javascript is disabled in this browser. This page requires Javascript. Modify your browser's settings to allow Javascript to execute. See your browser's documentation for specific instructions.
{}
Close
Publikační činnost
Probíhá načítání, čekejte prosím...
publicationId :
tempRecordId :
actionDispatchIndex :
navigationBranch :
pageMode :
tabSelected :
isRivValid :
Record type:
stať ve sborníku (D)
Home Department:
Ústav pro výzkum a aplikace fuzzy modelování (94410)
Title:
How image distortions affect inference accuracy
Citace
Dvořáček, P. How image distortions affect inference accuracy.
In:
2024 International Joint Conference on Neural Networks (IJCNN): 2024 International Joint Conference on Neural Networks (IJCNN) 2024-06-30 Yokohama.
Yokohama: IEEE, 2024. ISBN 979-8-3503-5931-2.
Subtitle
Publication year:
2024
Obor:
Number of pages:
7
Page from:
neuvedeno
Page to:
neuvedeno
Form of publication:
Elektronická verze
ISBN code:
979-8-3503-5931-2
ISSN code:
2161-4393
Proceedings title:
2024 International Joint Conference on Neural Networks (IJCNN)
Proceedings:
Mezinárodní
Publisher name:
IEEE
Place of publishing:
Yokohama
Country of Publication:
Sborník vydaný v zahraničí
Název konference:
2024 International Joint Conference on Neural Networks (IJCNN)
Místo konání konference:
Yokohama
Datum zahájení konference:
Typ akce podle státní
příslušnosti účastníků:
Celosvětová akce
WoS code:
EID:
2-s2.0-85204998496
Key words in English:
image processing, neural network inference, adversarial machine learning
Annotation in original language:
This research thoroughly analyses the effects of various image distortions on different neural network architectures, utilizing the ImageNet validation dataset. The study reveals that distortions such as blur, noise, JPEG compression, and grid patterns significantly elevate error rates. Contrast change yielded inconsistent impacts, only occasionally increasing error rates, while rescaling the image drastically (more than five times) led to substantial model performance degradation, although the images became difficult for even humans to recognize at such scales. Among the architectures tested, EfficientNet models demonstrated superior robustness, which can be attributed to their scalable input sizes. The Vision Transformer (ViT), trained on the extensive JFT-300M dataset, also showcased notable resilience to distortions. Other architectures like ConvNeXt, Swin, and ResNetD were more susceptible to the tested distortions. MobileNetV3-based models seem to provide an exceptionally good ratio between the robustness and model size, but this could also be caused by the fact that models from this family have inherently higher base error rates, which intrinsically allows for a larger margin of error.
Annotation in english language:
This research thoroughly analyses the effects of various image distortions on different neural network architectures, utilizing the ImageNet validation dataset. The study reveals that distortions such as blur, noise, JPEG compression, and grid patterns significantly elevate error rates. Contrast change yielded inconsistent impacts, only occasionally increasing error rates, while rescaling the image drastically (more than five times) led to substantial model performance degradation, although the images became difficult for even humans to recognize at such scales. Among the architectures tested, EfficientNet models demonstrated superior robustness, which can be attributed to their scalable input sizes. The Vision Transformer (ViT), trained on the extensive JFT-300M dataset, also showcased notable resilience to distortions. Other architectures like ConvNeXt, Swin, and ResNetD were more susceptible to the tested distortions. MobileNetV3-based models seem to provide an exceptionally good ratio between the robustness and model size, but this could also be caused by the fact that models from this family have inherently higher base error rates, which intrinsically allows for a larger margin of error.
References
Reference
R01:
RIV/61988987:17610/24:A25038FL
Complementary Content
Deferred Modules
${title}
${badge}
${loading}
Deferred Modules