deepPGSegNet: MRI-based pituitary gland segmentation using deep learning

Front Endocrinol (Lausanne). 2024 Feb 2:15:1338743. doi: 10.3389/fendo.2024.1338743. eCollection 2024.

Abstract

Introduction: In clinical research on pituitary disorders, pituitary gland (PG) segmentation plays a pivotal role, which impacts the diagnosis and treatment of conditions such as endocrine dysfunctions and visual impairments. Manual segmentation, which is the traditional method, is tedious and susceptible to inter-observer differences. Thus, this study introduces an automated solution, utilizing deep learning, for PG segmentation from magnetic resonance imaging (MRI).

Methods: A total of 153 university students were enrolled, and their MRI images were used to build a training dataset and ground truth data through manual segmentation of the PGs. A model was trained employing data augmentation and a three-dimensional U-Net architecture with a five-fold cross-validation. A predefined field of view was applied to highlight the PG region to optimize memory usage. The model's performance was tested on an independent dataset. The model's performance was tested on an independent dataset for evaluating accuracy, precision, recall, and an F1 score.

Results and discussion: The model achieved a training accuracy, precision, recall, and an F1 score of 92.7%, 0.87, 0.91, and 0.89, respectively. Moreover, the study explored the relationship between PG morphology and age using the model. The results indicated a significant association between PG volume and midsagittal area with age. These findings suggest that a precise volumetric PG analysis through an automated segmentation can greatly enhance diagnostic accuracy and surveillance of pituitary disorders.

Keywords: 3D UNet; MRI; deep learning; pituitary disorders; pituitary gland; segmentation.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Deep Learning*
  • Humans
  • Image Processing, Computer-Assisted / methods
  • Magnetic Resonance Imaging / methods
  • Pituitary Gland / diagnostic imaging
  • Research Design

Grants and funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This study was supported by JSPS KAKENHI Grant Number 19H00532 & 21H02800, a research grant from ‘Creative KMEDI hub’ in 2022. [No. B-C-N-22-10], and the Cooperative Study Program (23−633) of National Institute for Physiological Sciences.