Please use this identifier to cite or link to this item: https://hdl.handle.net/10356/17003
Title: Combining multiple image modalities for better image segmentation
Authors: Sari Setianingsih.
Keywords: DRNTU::Engineering::Computer science and engineering::Computing methodologies::Image processing and computer vision
Issue Date: 2009
Abstract: Accurate and robust brain/non-brain segmentation is very crucial in brain imaging application. Formerly, brain extraction relied on a single image modality, which limits its performance and accuracy. Nowadays, high resolution of T1- and T2-weighted images can be acquired during the same scanning session. This creates a promising possibility of combining images to improve delineation of brain structures. In this report, we present a novel skull striping algorithm which aims to get more accurate and robust extracted brain image region. The idea is by incorporating the information from T2-weigthed image into the skull striping decision process. In order to achieve this, the pair of images must be brought into strict correspondence. Perfect alignment is required. We also introduce a fresh approach on multi-modal image alignment. This is done by making use of the existing intra-modality image alignment. Hence, conversion from multimodal images into similar modality images is required. Thresholding approach is proposed to accomplish this conversion.
URI: http://hdl.handle.net/10356/17003
Schools: School of Computer Engineering 
Rights: Nanyang Technological University
Fulltext Permission: restricted
Fulltext Availability: With Fulltext
Appears in Collections:SCSE Student Reports (FYP/IA/PA/PI)

Files in This Item:
File Description SizeFormat 
SariSetianingsih09.pdf
  Restricted Access
1.68 MBAdobe PDFView/Open

Page view(s) 50

529
Updated on Mar 15, 2025

Download(s)

13
Updated on Mar 15, 2025

Google ScholarTM

Check

Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.