dc.contributor.author | He, Xiaonan | |
dc.contributor.author | Ye, Yuhang | |
dc.contributor.author | Lee, Brian | |
dc.contributor.author | Xia, Yukun | |
dc.contributor.author | Qiao, Yuansong | |
dc.date.accessioned | 2025-01-07T12:28:07Z | |
dc.date.available | 2025-01-07T12:28:07Z | |
dc.date.issued | 2024-07-02 | |
dc.identifier.citation | He, X., Ye, Y., Lee, B., Xia, Y. and Qiao, Y. (2024) 'Semantic guidance incremental network for efficiency video super-resolution', Visual Computer, 40(7), pp. 4899-4911. Available at: DOI: 10.1007/s00371-024-03488-y | en_US |
dc.identifier.uri | https://research.thea.ie/handle/20.500.12065/4879 | |
dc.description.abstract | In video streaming, bandwidth constraints significantly affect client-side video quality. Addressing this, deep neural networks offer a promising avenue for implementing video super-resolution (VSR) at the user end, leveraging advancements in modern hardware, including mobile devices. The principal challenge in VSR is the computational intensity involved in processing temporal/spatial video data. Conventional methods, uniformly processing entire scenes, often result in inefficient resource allocation. This is evident in the over-processing of simpler regions and insufficient attention to complex regions, leading to edge artifacts in merged regions. Our innovative approach employs semantic segmentation and spatial frequency-based categorization to divide each video frame into regions of varying complexity: simple, medium, and complex. These are then processed through an efficient incremental model, optimizing computational resources. A key innovation is the sparse temporal/spatial feature transformation layer, which mitigates edge artifacts and ensures seamless integration of regional features, enhancing the naturalness of the super-resolution outcome. Experimental results demonstrate that our method significantly boosts VSR efficiency while maintaining effectiveness. This marks a notable advancement in streaming video technology, optimizing video quality with reduced computational demands. This approach, featuring semantic segmentation, spatial frequency analysis, and an incremental network structure, represents a substantial improvement over traditional VSR methodologies, addressing the core challenges of efficiency and quality in high-resolution video streaming. | en_US |
dc.format | application/pdf | en_US |
dc.language.iso | eng | en_US |
dc.publisher | Springer Nature | en_US |
dc.relation.ispartof | Visual Computer | en_US |
dc.rights | Attribution 4.0 International | * |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | * |
dc.subject | Convolutional neural network | en_US |
dc.subject | Efficiency | en_US |
dc.subject | Semantic guidance | en_US |
dc.subject | Video super-resolution | en_US |
dc.title | Semantic guidance incremental network for efficiency video super-resolution | en_US |
dc.type | info:eu-repo/semantics/article | en_US |
dc.contributor.affiliation | Technological University of the Shannon: Midlands Midwest | en_US |
dc.description.peerreview | yes | en_US |
dc.identifier.doi | 10.1007/s00371-024-03488-y | en_US |
dc.identifier.eissn | 1432-2315 | |
dc.identifier.endpage | 4911 | en_US |
dc.identifier.issue | 7 | en_US |
dc.identifier.orcid | https://orcid.org/0000-0002-8475-4074 | en_US |
dc.identifier.startpage | 4899 | en_US |
dc.identifier.volume | 40 | en_US |
dc.rights.accessrights | info:eu-repo/semantics/openAccess | en_US |
dc.subject.department | Department of Computer and Software Engineering | en_US |
dc.type.version | info:eu-repo/semantics/publishedVersion | en_US |