Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/184038
Title: | SA-LUT: spatial adaptive 4D look-up table for photorealistic style transfer | Authors: | Gong, Zerui | Keywords: | Computer and Information Science | Issue Date: | 2025 | Publisher: | Nanyang Technological University | Source: | Gong, Z. (2025). SA-LUT: spatial adaptive 4D look-up table for photorealistic style transfer. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/184038 | Abstract: | With the rapid advancement of technology, photorealistic style transfer (PST) has become increasingly significant in film post-production and professional photography, owing to its capability to adapt the color characteristics of a reference image while preserving the structural content. However, current methods exhibit notable limitations for practical implementation. Generation-based techniques often prioritize stylistic fidelity at the expense of content integrity and computational efficiency, whereas global color transformation methods, such as those based on look-up tables (LUTs), maintain structural consistency but lack spatial adaptability. To address these limitations, this thesis presents the Spatial Adaptive 4D Look-Up Table (SA-LUT), a novel framework that integrates the efficiency of LUTs with the adaptability of neural networks. The proposed approach comprises two key compo- nents: a Style-guided 4D LUT Generator, which extracts multi-scale features from the style image to predict a 4D LUT; and a Context Generator, which utilizes content-style cross-attention to produce a context map for spatially adaptive adjustment. This context map enables precise, location-specific color transformations while preserving structural integrity. Furthermore, this thesis introduces PST50, the first benchmark dataset specifically designed for the evaluation of photorealistic style transfer performance. Experimental results demonstrate that the proposed SA-LUT framework significantly outperforms existing state-of-the-art methods, achieving a 66.7% reduction in perceptual distance relative to 3D LUT-based approaches, while maintaining real-time video stylization performance at 16 frames per second. | URI: | https://hdl.handle.net/10356/184038 | Schools: | College of Computing and Data Science | Fulltext Permission: | restricted | Fulltext Availability: | With Fulltext |
Appears in Collections: | CCDS Student Reports (FYP/IA/PA/PI) |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
Gong_Zerui_FYP-Submission.pdf Restricted Access | 39.75 MB | Adobe PDF | View/Open |
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.