Skip to main content
Advertisement

Main menu

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • ASNR Foundation Special Collection
    • Most Impactful AJNR Articles
    • Photon-Counting CT
    • Spinal CSF Leak Articles (Jan 2020-June 2024)
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home

User menu

  • Alerts
  • Log in

Search

  • Advanced search
American Journal of Neuroradiology
American Journal of Neuroradiology

American Journal of Neuroradiology

ASHNR American Society of Functional Neuroradiology ASHNR American Society of Pediatric Neuroradiology ASSR
  • Alerts
  • Log in

Advanced Search

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • ASNR Foundation Special Collection
    • Most Impactful AJNR Articles
    • Photon-Counting CT
    • Spinal CSF Leak Articles (Jan 2020-June 2024)
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home
  • Follow AJNR on Twitter
  • Visit AJNR on Facebook
  • Follow AJNR on Instagram
  • Join AJNR on LinkedIn
  • RSS Feeds

AJNR is seeking candidates for the AJNR Podcast Editor. Read the position description.

Research ArticleSpine

Deep Learning–Generated Synthetic MR Imaging STIR Spine Images Are Superior in Image Quality and Diagnostically Equivalent to Conventional STIR: A Multicenter, Multireader Trial

L.N. Tanenbaum, S.C. Bash, G. Zaharchuk, A. Shankaranarayanan, R. Chamberlain, M. Wintermark, C. Beaulieu, M. Novick and L. Wang
American Journal of Neuroradiology August 2023, 44 (8) 987-993; DOI: https://doi.org/10.3174/ajnr.A7920
L.N. Tanenbaum
aFrom RadNet (L.N.T., S.C.B.), New York, New York
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for L.N. Tanenbaum
S.C. Bash
aFrom RadNet (L.N.T., S.C.B.), New York, New York
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for S.C. Bash
G. Zaharchuk
bStanford University Medical Center (G.Z., C.B.), Stanford, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for G. Zaharchuk
A. Shankaranarayanan
cSubtle Medical (A.S., R.C., L.W.), Menlo Park, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for A. Shankaranarayanan
R. Chamberlain
cSubtle Medical (A.S., R.C., L.W.), Menlo Park, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for R. Chamberlain
M. Wintermark
dMD Anderson Cancer Center (M.W.), University of Texas, Houston, Texas
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for M. Wintermark
C. Beaulieu
bStanford University Medical Center (G.Z., C.B.), Stanford, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for C. Beaulieu
M. Novick
eAll-American Teleradiology (M.N.), Bay Village, Ohio
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
L. Wang
cSubtle Medical (A.S., R.C., L.W.), Menlo Park, California
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for L. Wang
  • Article
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF
Loading

Abstract

BACKGROUND AND PURPOSE: Deep learning image reconstruction allows faster MR imaging acquisitions while matching or exceeding the standard of care and can create synthetic images from existing data sets. This multicenter, multireader spine study evaluated the performance of synthetically created STIR compared with acquired STIR.

MATERIALS AND METHODS: From a multicenter, multiscanner data base of 328 clinical cases, a nonreader neuroradiologist randomly selected 110 spine MR imaging studies in 93 patients (sagittal T1, T2, and STIR) and classified them into 5 categories of disease and healthy. A DICOM-based deep learning application generated a synthetically created STIR series from the sagittal T1 and T2 images. Five radiologists (3 neuroradiologists, 1 musculoskeletal radiologist, and 1 general radiologist) rated the STIR quality and classified disease pathology (study 1, n = 80). They then assessed the presence or absence of findings typically evaluated with STIR in patients with trauma (study 2, n = 30). The readers evaluated studies with either acquired STIR or synthetically created STIR in a blinded and randomized fashion with a 1-month washout period. The interchangeability of acquired STIR and synthetically created STIR was assessed using a noninferiority threshold of 10%.

RESULTS: For classification, there was a decrease in interreader agreement expected by randomly introducing synthetically created STIR of 3.23%. For trauma, there was an overall increase in interreader agreement by +1.9%. The lower bound of confidence for both exceeded the noninferiority threshold, indicating interchangeability of synthetically created STIR with acquired STIR. Both the Wilcoxon signed-rank and t tests showed higher image-quality scores for synthetically created STIR over acquired STIR (P < .0001).

CONCLUSIONS: Synthetically created STIR spine MR images were diagnostically interchangeable with acquired STIR, while providing significantly higher image quality, suggesting routine clinical practice potential.

ABBREVIATIONS:

Acq-STIR
acquired STIR
CNN
convolutional neural network
DL
deep learning
IQ
image quality
RMSE
root mean square error
RMSPE
root mean square percentage error
Syn-STIR
synthetically created STIR

A typical clinical protocol for spine MR imaging uses T1WI, T2WI, and STIR scans to depict anatomy and provide adequate sensitivity to a variety of pathologic conditions.

STIR offers a combination of T1 and T2 contrast-weighting and nulled fat signal to highlight pathologic changes in tissues. The fat suppression offered by STIR is more uniform and resistant to magnetic field inhomogeneities than other fat-saturation methods such as spectral “fat-sat,” especially near metallic foreign bodies, tissue interfaces with high susceptibility differences (like the skull base/sinuses), and across large body parts like the spine. On lower-field permanent magnets with lower homogeneity, STIR may be the only fat-suppression method available. STIR images have inherently lower SNR than T1WI and T2WI. Despite approaches that use larger voxel sizes to mitigate this challenge, scan times are still long, more susceptible to motion, and harder for patients to tolerate.

Deep learning (DL)-based reconstruction techniques mitigate this challenge by enabling faster acquisitions while matching or even exceeding standard-of-care image quality (IQ).1⇓⇓-4 Recent work has led to DL methods that can generate entirely synthetic image contrasts, potentially shortening overall study times by removing the need to acquire certain series.

Synthesizing new contrast information from available images has been an active area of research in the MR imaging domain. Convolutional neural network (CNN)-based approaches have demonstrated state-of-the-art performance for MR imaging contrast synthesis.5⇓⇓-8 While most of the literature has focused on one-to-one synthesis, several studies considered the many-to-one synthesis problem, in which the algorithm takes multiple contrasts as input and generates 1 missing contrast.5,7,9 Previous work has demonstrated the potential to synthesize STIR images from T1WI and T2WI.10 This study goes further by using examinations from multiple scanner manufacturers and a wider variety of magnetic field strengths, including a more comprehensive set of pathologies, and by performing voxelwise analysis of the synthetically created STIR (Syn-STIR) images. Furthermore, the technical methods used in this work integrate multiple maps (including an anatomy-aware segmentation map and a pathology saliency map) in the reconstruction network, so the syn-STIR images maintain high consistency with the acquired STIR (Acq-STIR) images. Our methods also avoid the use of generative adversarial networks, which are prone to introducing structures in synthesized images that are not present in the source images.

This multicenter, multireader study evaluated the diagnostic interchangeability and qualitative image quality of a DL-generated Syn-STIR against a clinical standard-of-care Acq-STIR. There are established methods to assess the interchangeability of the 2 image-acquisition methods, which determine whether the images are diagnostically equivalent.11 Two images are interchangeable or diagnostically equivalent if a given patient would receive the same diagnosis regardless of which of the 2 images was used. Diagnostic equivalence is tested by comparing an interreader agreement using the baseline imaging method with an interreader agreement using the method being tested versus the baseline while accounting for variability across cases and readers. In addition, multiple quantitative methods were also used to compare the Syn-STIR images with the Acq-STIR images.

MATERIALS AND METHODS

Overview

A DL model was applied to synthesize a sagittal Syn-STIR series from the sagittal T1 and sagittal T2 of clinical spine MR imaging studies. The model contains 3 phases, an anatomy-aware map, a pathology-aware map, and a reconstruction map. For the anatomy-aware map, the segmentation map was obtained for each anatomy, making the anatomy-based operation feasible. The pathology-aware map is a saliency map used to guide the network to maintain pathologic consistency. During the training process, the 2 inputs (sagittal T1 and sagittal T2) were fed into the reconstruction network under separate branches and later concatenated to avoid potential blurriness due to misregistration. We implemented the DL model in TensorFlow (https://www.tensorflow.org/), trained on an NVIDIA V100 GPU (https://www.nvidia.com/en-us/data-center/v100/) with an ADAM optimizer (https://machinelearningjourney.com/index.php/2021/01/09/adam-optimizer/),12 and applied image registration between the 2 inputs to reduce potential misalignment. The network was trained by comparing the output Syn-STIR image with the Acq-STIR image through multiple loss functions (Online Supplemental Data).

Participants and Distribution of Pathologies

With institutional review board approval, a nonreader senior neuroradiologist identified the dominant pathology (as described below) in a multicenter, multiscanner data base of 328 approximately equal numbers of cervical, thoracic, and lumbar spine MR imaging cases referred for a variety of conditions. From this group, 93 unique patients were evaluated in 2 separate studies. First, 80 patients (40 females, 36 males, 4 not available; age range, 16–89 years) were selected randomly from among 5 categories of disease (defined as the most dominant pathology) based on the findings on the complete study (study 1). The categories were cord lesion (n = 8), noncord lesion (n = 15), degenerative disease (n = 20), infection (n = 10), trauma (n = 17), and healthy (n = 10). The readers were given instructions outlining which clinical entities should fall into each category and to help with classification when multiple pathologies were present. More details can be found in the Online Supplemental Data. In addition, a second study evaluating the ability of readers to identify important features in the setting of trauma was performed (study 2). Patients (13 men, 17 women; age range, 18–89 years) for study 2 included 10 with no imaging evidence of trauma (separate from the patients in study 1) and 20 with imaging evidence of trauma (17 of the patients with trauma in study 1 supplemented by 3 additional patients). These cases were evaluated for the following findings: prevertebral fluid collections (class I), bone edema related to fracture (class II), and posterior soft-tissue/ligamentous injury (class III). The case distribution was class I/II (n = 3), class I/II/III (n = 7), class I/III (n = 5), class II (n = 1), class II/III (n = 1), class III (n = 4), and class none (n = 9).

Image Acquisition

The images were acquired on a variety of scanners, including 3T Discovery 750 and 750w, 3T Signa Premier, and 1.5T HDxt (GE Healthcare); 3T Magnetom Skyra, 3T MagnetomVerio (Siemens); 1.5T Intera (Philips Healthcare); 1.5T Vantage Titan (Canon); 0.6T (Fonar Upright) and 0.3T AIRIS Elite (Hitachi/Fujifilm). The case distribution by field strength for study 1 was 0.6T (n = 1), 1T (n = 1), 1.5T (n = 43), and 3T (n = 35), and for study 2, it was 1T (n = 2), 1.5T (n = 16), and 3T (n = 12). The image acquisitions consisted of sagittal T1, T2, and STIR series using the individual institution’s routine clinical protocol. Section thickness ranged from 3 to 5 mm. FOV varied from 18 to 24 cm cervical; 27 to 30 cm lumbar; and 30 to 38 cm thoracic; and the acquisition matrix varied from 192 × 192 to 800 × 380.

Image Processing

The Syn-STIR images were created off-line from existing DICOM images using a vendor-neutral, CNN software application (SubtleSYNTH; Subtle Medical). The CNN was trained to generate synthetic sagittal STIR images using the sagittal T1 and T2 images as input. Because the application was DICOM-based, processing did not require proprietary raw k-space input; thus, it was capable of processing images from any MR imaging platform. The training set included hundreds of thousands of MR images from a variety of vendors, scanner models, field strengths, and clinical sites, as well as a variety of disease states/clinical indications, thus experiencing a range of tissue contrasts, acquisition parameters, patient anatomies, and image quality.

Image Assessment

Five radiologists (3 neuroradiologists, 1 musculoskeletal radiologist, 1 general radiologist experienced in spine MR imaging) evaluated 160 cases. Each case consisted of 3 sagittal image series: either T1, T2, and Acq-STIR (n = 80) or T1, T2, and Syn-STIR (n = 80). The image sets were presented in a blinded and randomized fashion on a commercial DICOM viewer, with a 1-month washout period between reading sessions (study 1). To assess diagnostic equivalence, each reader individually classified the pathologies present. Readers also rated the Acq-STIR and Syn-STIR image quality on a 5-point Likert scale (1 = unacceptable, 2 = poor, 3 = adequate, 4 = good, 5 = excellent), which served as a collective summary assessment of individual image-quality metrics, such as perceived SNR, contrast-to-noise ratio, image sharpness, and artifacts. The same readers also evaluated the trauma-specific study (n = 30 subjects, n = 60 studies) in the same blinded and randomized fashion with the same 1-month washout period (study 2). They were asked to individually classify the findings for the presence/absence of the following: 1) prevertebral fluid collections, 2) fracture-related bone edema, and 3) posterior soft-tissue/ligamentous injury. In addition, 2 neuroradiologists, including one not involved with studies 1 or 2, performed a blinded side-by-side, qualitative evaluation of the study 1 Acq-STIR and Syn-STIR images, rating the extent of disease and diagnostic confidence on a 5-point Likert scale, as well as noting evidence or absence of image aberrations.

RESULTS

Diagnostic-Equivalence Analysis

The 2 imaging methods were assessed for interchangeability or diagnostic equivalence by comparing the interreader agreement within Acq-STIR images with the interreader agreement between Syn-STIR and Acq-STIR images. Interreader agreement for Acq-STIR images was calculated as the percentage of comparisons between 2 different readers for the same case in which the readers’ classifications agreed. Interreader agreement for Acq-STIR versus Syn-STIR images was calculated as the percentage of comparisons between 2 different readers for the same case that the readers’ classification when 1 reader was using the Acq-STIR image and the other reader was using the Syn-STIR image agreed with each other. The agreement probability was calculated by mean of a logistic regression model with random effects using the “glmer” function from the “lme4” package in R statistical and computing software (http://www.r-project.org/) following methods described in the literature.11,13 A noninferiority analysis was performed with a preset hypothesis that the difference in diagnostic classification for interreader agreement for Acq-STIR and Syn-STIR was not >10% lower than the interreader agreement between Syn-STIR and Acq-STIR.

Image-Quality Statistical Analysis

Wilcoxon rank-sum tests were performed to assess the equivalence or superiority of the image quality for each feature. Statistically significant superiority for a feature was determined by P < .05. Adjustment for significance tests for multiple comparisons was made using a Bonferroni correction.

Voxel-Intensity Analysis

To evaluate the voxelwise correlation between the Syn-STIR image and the conventional STIR image, we drew 4 ROIs on each target tissue (vertebral bone, disc, CSF, spinal cord, and fat) and calculated the mean of the 4 ROIs per series. Note that areas without any pathologies were selected. For example, in patients with degenerative disease, ROIs were drawn only on the healthy disc. Similar rules applied to other tissues as well. Bland-Altman analysis14,15 was then applied, followed by the Shapiro-Wilk results on the difference.16 Additionally, root mean square error (RMSE) and root mean square percentage error (RMSPE) were calculated. A Passing-Bablok regression analysis was performed to evaluate agreement between the 2 images.17

Sample image pairs are shown in Fig 1, demonstrating similar fat-saturated T2-weighted image contrast of the Syn-STIR as the Acq-STIR. Lower noise levels are seen in the Syn-STIR images.

FIG 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 1.

A, L2 body fracture. Conventional acquired STIR (left) and DL synthesized STIR (right). Note the Syn-STIR manifests fewer motion artifacts, improved sharpness, and greater apparent SNR. B, Sacral metastasis. Conventional Acq-STIR (left) and DL Syn-STIR (right). Note the striking reduction in motion artifacts as well as superior overall quality of the Syn-STIR.

Diagnostic Interchangeability

The estimate of interchangeability (diagnostic equivalence) when accounting for readers and cases as random effects was −3.23% (95% CI, –6.61%–0.19%), evaluated over 1000 bootstrapped samples (Fig 2). The decrease in interreader agreement expected when interchanging Acq-STIR images with Syn-STIR images was 3.23%. Based on the results, the estimate of interchangeability was not significantly worse than the noninferiority limit of 10% (P = .001). On the basis of the prespecified noninferiority criteria of 10%, we concluded that interchanging the Acq-STIR images with Syn-STIR images would not lead to a significant decrease in interreader agreement; thus, the Syn-STIR was deemed diagnostically equivalent to the Acq-STIR.

FIG 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 2.

Interchangeability when accounting for readers and cases as random effects under the whole-cohort estimated >1000 bootstrapped samples.

For the trauma subset, the 3 structure-based classifications (prevertebral fluid collections, fracture-related bone edema, and posterior soft-tissue/ligamentous injury) were analyzed separately as different classes (Fig 3). The estimate of interchangeability (diagnostic equivalence) when accounting for readers and cases as random effects for the 3 classes was +0.85% (95% CI, –4.13%–5.48%), +2.3% (95% CI, –2.8%–7.1%), and +2.2% (95% CI, –2.2%–6.4%), respectively; each class evaluated >1000 bootstrapped samples. In other words, the interreader agreement can be expected to improve by 0.85%, 2.3%, and 2.2% when interchanging traditional STIR images with Syn-STIR images. Based on the results, the estimate of interchangeability was not significantly worse than the noninferiority limit of 10% (P = .001). Given all 3 classes, a final analysis was performed in which the results described above from the 3 classes were combined, and “class” was included as a fixed effect in the statistical model. The interchangeability estimate was +1.9% (95% CI, –1.1%–5.0%), indicating that there was an improvement in the interreader agreement found when interchanging Acq-STIR images with Syn-STIR images. We, therefore, conclude that for the trauma study Syn-STIR was interchangeable with Acq-STIR.

FIG 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 3.

Interchangeability when accounting for readers and cases as random effects under the trauma cohort estimated >1000 bootstrapped samples. The x-axis represents the estimated interchangeabilities in the unit of percentage, and the y-axis represents the count of bootstrapped samples. A, Trauma class I. B, Trauma class II. C, Trauma class III. D, Mixed effect for all 3 classes.

Image-Quality Analysis

Acq-STIR images had an average IQ score of 3.21 (SD, 1.08), and Syn-STIR images scored an average of 3.71 (SD, 1.14). A Wilcoxon signed-rank test showed a significantly higher median IQ score for Syn-STIR images than Acq-STIR images (median = 0.4, P < .0001). A t test on the paired difference in IQ scores across artificial intelligence–generated and Acq-STIR images showed a significantly higher average IQ score for Syn-STIR images compared with Acq-STIR images (mean paired difference = 0.50; 95% CI, 0.33–0.67; P < .0001).

Side-by-Side Comparison

In the blinded, side-by-side evaluation of the cases in study one, 94.9% of Syn-STIR sets demonstrated equal or a better extent of disease compared with Acq-STIR for the first reader and 97.5% for the second reader; 88.6% of cases provided equal or higher diagnostic confidence with Syn-STIR for the first reader and 87.3% for the second reader. In addition, no unexpected differences were found between the 2 STIR types, indicating that the Syn-STIR method did not create unique artifacts.

Voxel Consistency

The Bland-Altman plots for voxel consistency are shown in Fig 4. For each tissue, the bias (the mean of the difference between the Acq-STIR and Syn-STIR) was close to zero. The smallest average bias was from the CSF, which was −0.04 normalized intensity units, and the largest average bias was from fat, which was around −0.25 normalized intensity units. The Shapiro-Wilk results showed that all P values were > .05, implying that the difference between the Acq-STIR and the Syn-STIR is normally distributed.

FIG 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 4.

Bland-Altman plots for each tissue from 80 clinical cases. The x-axis represents the mean of the normalized intensity value from Syn-STIR and its nearest neighbor from Acq-STIR, and the y-axis represents the difference between the normalized intensity value from Syn-STIR and its nearest neighbor from the Acq-STIR.

In addition, the RMSE and RMSPE between the Acq-STIR and Syn-STIR images for each patient were 0.45 and 17.88 normalized intensity units, respectively. For all 80 cases, the median of the RMSE value was 0.45 normalized intensity units, and the median of the RMSPE percentage was 17.9%. After confirming that the 5 tissues passed the Shapiro-Wilk test for normality, the Passing-Bablok regression was applied to estimate the regression line and intercept (Fig 5). The slopes of the disc, CSF, and spinal cord were 1.06, 1.05, and 1.07, which indicate a high correlation between the 2 results. The slope of bone and fat was 0.85 and 0.78, respectively. The results indicate excellent voxelwise consistency between the Acq-STIR and the Syn-STIR images.

FIG 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 5.

Passing-Bablok regression applied to estimate the regression line and intercept. The slope of the disc, CSF, and spinal cord is 1.06, 1.05, and 1.07, which indicates a high correlation between the 2 results. The slope of bone and fat is 0.85 and 0.78, respectively. The results indicate excellent voxelwise consistency between the Acq-STIR and Syn-STIR images.

DISCUSSION

STIR is quite powerful in depicting spine pathology and thus is part of almost all routine spine imaging protocols; however, conventional reconstruction scan times are long. Also, because of the fat inversion pulse, the SNR of the images is lower than that of other sequences. A synthetically generated STIR could result in approximately 3–5 minutes of scan time avoided or up to 25% overall time-savings per examination, increasing imaging enterprise efficiency. Because up to 30% of patients report significant anxiety, largely from claustrophobia, during an MR imaging study, scan-time reductions inherently improve the patient’s experience.18 The authors’ internal multicenter surveys have shown that even minor reductions in examination length result in a significantly higher level of patient satisfaction.19

MR imaging examinations are susceptible to image degradation from motion, particularly during lengthy scans. Motion is a significant challenge in MR imaging, occurring in 29% of inpatient/emergency department examinations and 7% of outpatient studies20 and can lead to the need to repeat sequences or entire studies. Andre et al21 found that 19.8% of all MR imaging sequences needed to be repeated due to motion artifacts, correlating with US $592 revenue loss per hour and an annual loss of US $115,000 per scanner.

The generally inverse relationship between MR image quality and scan duration is well-established.1,2 Traditionally reconstructed, high-resolution, high-SNR images require acquisition times that can be quite long. DL-based image reconstruction is increasingly used in practice to reduce the time required to provide high-quality images by up to 50%.1,2 DL image synthesis offers effective 100% series acceleration. In addition, because the synthesized images in our study receive the SNR and spatial resolution of the acquired T1WI and T2WI scans, Syn-STIR can be expected to offer better image quality than is practical with an Acq-STIR.

Previous work on MR imaging sequence-to-sequence translation has been performed4⇓⇓⇓⇓⇓-10 but generally in subjects without pathology. This study demonstrated excellent performance in a patient cohort with a diverse set of typical spinal pathologies and evaluated key imaging findings commonly assessed with STIR imaging.

Absolute quality ratings could potentially obscure subtle failures and artifacts in the synthetically reconstructed image. Thus, a blinded, side-by-side evaluation was performed to compare the extent of disease and diagnostic confidence as well as to interrogate for evidence of image aberrations. We found that 96% of Syn-STIR sets manifested equal or better extent of disease compared with Acq-STIR, and 88% of cases provided equal or higher diagnostic confidence with Syn-STIR. Most important, no unexpected image appearances (“hallucinations”) or information losses were detected. Although our study had no cases in which the network failed to generate an acceptable Syn-STIR image, the quality of the Syn-STIR image depends on the quality of the input T1 and T2 images. Therefore, if the input images were to have gross artifacts or high noise levels, these could manifest on the Syn-STIR series.

Our study patients were imaged on scanners of differing vendors and field strengths, drawn from a variety of geographically diverse facilities, and encompassed a variety of disease entities, but we acknowledge a risk of inadvertent patient-selection bias or disease-representation bias during the initial gathering of the larger patient cohort.

In this randomized, blinded trial, Syn-STIR demonstrated superior image quality with respect to Acq-STIR. A potential limitation is that the overall qualitative image-quality assessment was a collective summary of perceived metrics against typical expectations and is thus biased by subjective preferences. However, quantitative measures, such as statistical analysis of voxel consistency across STIR data sets, were robust. Future exploration could apply synthetic image generation to additional body parts and other scanning techniques.

CONCLUSIONS

DL-based Syn-STIR MR images, derived from acquired T1WI and T2WI DICOM data sets from multiple centers, scanners, and field strengths, proved statistically interchangeable in diagnostic performance with traditionally acquired STIR and provided superior perceived image quality. Quantitative measures demonstrated consistent results, validating both the high accuracy of the Syn-STIR images and the generalizability of the DL method. This Syn-STIR method offers a promising clinical solution for faster and more comfortable spine MR imaging examinations.

Footnotes

  • Disclosure forms provided by the authors are available with the full text and PDF of this article at www.ajnr.org.

References

  1. 1.↵
    1. Bash S,
    2. Johnson B,
    3. Gibbs W, et al
    . Deep learning image processing enables 40% faster spinal MR scans which match or exceed quality of standard of care. Clin Neuroradiol 2022;32:197–203 doi:10.1007/s00062-021-01121-2 pmid:34846555
    CrossRefPubMed
  2. 2.↵
    1. Bash S,
    2. Wang L,
    3. Airriess C, et al
    . Deep learning enables 60% accelerated volumetric brain MRI while preserving quantitative performance: a prospective, multicenter, multireader trial. AJNR Am J Neuroradiol 2021;42:2130–37 doi:10.3174/ajnr.A7358 pmid:34824098
    Abstract/FREE Full Text
  3. 3.↵
    1. Zaharchuk G,
    2. Gong E,
    3. Wintermark M, et al
    . Deep learning in neuroradiology. AJNR Am J Neuroradiol 2018;39:1776–84 doi:10.3174/ajnr.A5543 pmid:29419402
    Abstract/FREE Full Text
  4. 4.↵
    1. Liu J,
    2. Pasumarthi S,
    3. Duffy B, et al
    . One model to synthesize them all: multi-contrast multi-scale transformer for missing data imputation. IEEE Trans Med Imaging 2023 Mar 4 [Epub ahead of print] doi:10.1109/TMI.2023.3261707 pmid:37030684
    CrossRefPubMed
  5. 5.↵
    1. Shen E,
    2. Liu T,
    3. Peters, TM
    1. Li H,
    2. Paetzold JC,
    3. Sekuboyina A, et al
    . DiamondGAN: unified multi-modal generative adversarial networks for MRI sequences synthesis. In: Shen E, Liu T, Peters, TM, et al. Medical Image Computing and Computer Assisted Intervention–MICCAI. Springer International Publishing; 2019:795–803
  6. 6.↵
    1. Denck J,
    2. Guehring J,
    3. Maier A, et al
    . Enhanced magnetic resonance image synthesis with contrast-aware generative adversarial networks. J Imaging 2021;7:133 doi:10.3390/jimaging7080133 pmid:34460769
    CrossRefPubMed
  7. 7.↵
    1. Yurt M,
    2. Dar SU,
    3. Erdem A, et al
    . mustGAN: multi-stream Generative Adversarial Networks for MR image synthesis. Med Image Anal 2021;70:101944 doi:10.1016/j.media.2020.101944 pmid:33690024
    CrossRefPubMed
  8. 8.↵
    1. Dar SU,
    2. Yurt M,
    3. Karacan L, et al
    . Image synthesis in multi-contrast MRI with conditional generative adversarial networks. IEEE Trans Med Imaging 2019;38:2375–88 doi:10.1109/TMI.2019.2901750 pmid:30835216
    CrossRefPubMed
  9. 9.↵
    1. Lee D,
    2. Kim J,
    3. Moon WJ, et al
    . CollaGAN: Collaborative GAN for Missing Image Data Imputation. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, Long Beach, California. June 15–20, 2019:2487–96 doi:10.1109/CVPR.2019.00259
    CrossRef
  10. 10.↵
    1. Haubold J,
    2. Demircioglu A,
    3. Theysohn JM, et al
    . Generating virtual short tau inversion recovery (STIR) images from T1- and T2-weighted images using a conditional generative adversarial network in spine imaging. Diagnostics (Basel) 2021;11:1542 doi:10.3390/diagnostics11091542 pmid:34573884
    CrossRefPubMed
  11. 11.↵
    1. Obuchowski NA,
    2. Subhas N,
    3. Schoenhagen P
    . Testing for interchangeability of imaging tests. Acad Radiol 2014;21:1483–89 doi:10.1016/j.acra.2014.07.004 pmid:25300725
    CrossRefPubMed
  12. 12.↵
    1. Kingma DP,
    2. Ba J
    . Adam: A Method for Stochastic Optimization. arXiv. December 22, 2014. http://arxiv.org/abs/1412.6980. Accessed January 15, 2023
  13. 13.↵
    1. Khatri G,
    2. Pedrosa I,
    3. Ananthakrishnan L, et al
    . Abbreviated-protocol screening MRI vs. complete-protocol diagnostic MRI for detection of hepatocellular carcinoma in patients with cirrhosis: an equivalence study using LI-RADS v2018. J Magn Reson Imaging 2020;51:415–25 doi:10.1002/jmri.26835 pmid:31209978
    CrossRefPubMed
  14. 14.↵
    1. Giavarina D
    . Understanding Bland Altman analysis. Biochem Med (Zagreb) 2015;25:141–51 doi:10.11613/BM.2015.015 pmid:26110027
    CrossRefPubMed
  15. 15.↵
    1. Altman DG,
    2. Bland JM
    . Measurement in medicine: the analysis of method comparison studies. The Statistician 1983;32:307–17
    CrossRef
  16. 16.↵
    1. Henderson AR
    . Testing experimental data for univariate normality. Clin Chim Acta 2006;366:112–29 doi:10.1016/j.cca.2005.11.007 pmid:16388793
    CrossRefPubMed
  17. 17.↵
    1. Sotgia S,
    2. Zinellu A,
    3. Pinna GA, et al
    . A new general regression-based approach for method comparison studies. Clin Chem Lab Med 2008;46:1046–49 doi:10.1515/CCLM.2008.194 pmid:18605944
    CrossRefPubMed
  18. 18.↵
    1. Tanenbaum L
    . Quality, efficiency, and survival with patient centric imaging. In: Proceedings of 7th Annual Snowmass 2023: Hot Topics in Radiology: Advanced Imaging and Artificial Intelligence. Snowmass Village, Colorado. February 5–10, 2019
  19. 19.↵
    1. Meléndez JC,
    2. McCrank E
    . Anxiety-related reactions associated with magnetic resonance imaging examinations. JAMA 1993;270:745–47 doi:10.1001/jama.1993.03510060091039 pmid:8336378
    CrossRefPubMed
  20. 20.↵
    1. Zaitsev M,
    2. Maclaren J,
    3. Herbst M
    . Motion artifacts in MRI: a complex problem with many partial solutions. J Magn Reson Imaging 2015;42:887–901 doi:10.1002/jmri.24850 pmid:25630632
    CrossRefPubMed
  21. 21.↵
    1. Andre JB,
    2. Bresnahan BW,
    3. Mossa-Basha M, et al
    . Toward quantifying the prevalence, severity, and cost associated with patient motion during clinical MR examinations. J Am Coll Radiol 2015;12:689–95 doi:10.1016/j.jacr.2015.03.007 pmid:25963225
    CrossRefPubMed
  • Received March 20, 2023.
  • Accepted after revision June 1, 2023.
  • © 2023 by American Journal of Neuroradiology
PreviousNext
Back to top

In this issue

American Journal of Neuroradiology: 44 (8)
American Journal of Neuroradiology
Vol. 44, Issue 8
1 Aug 2023
  • Table of Contents
  • Index by author
  • Complete Issue (PDF)
Advertisement
Print
Download PDF
Email Article

Thank you for your interest in spreading the word on American Journal of Neuroradiology.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Deep Learning–Generated Synthetic MR Imaging STIR Spine Images Are Superior in Image Quality and Diagnostically Equivalent to Conventional STIR: A Multicenter, Multireader Trial
(Your Name) has sent you a message from American Journal of Neuroradiology
(Your Name) thought you would like to see the American Journal of Neuroradiology web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Cite this article
L.N. Tanenbaum, S.C. Bash, G. Zaharchuk, A. Shankaranarayanan, R. Chamberlain, M. Wintermark, C. Beaulieu, M. Novick, L. Wang
Deep Learning–Generated Synthetic MR Imaging STIR Spine Images Are Superior in Image Quality and Diagnostically Equivalent to Conventional STIR: A Multicenter, Multireader Trial
American Journal of Neuroradiology Aug 2023, 44 (8) 987-993; DOI: 10.3174/ajnr.A7920

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
0 Responses
Respond to this article
Share
Bookmark this article
Deep Learning for STIR Spine MRI Quality
L.N. Tanenbaum, S.C. Bash, G. Zaharchuk, A. Shankaranarayanan, R. Chamberlain, M. Wintermark, C. Beaulieu, M. Novick, L. Wang
American Journal of Neuroradiology Aug 2023, 44 (8) 987-993; DOI: 10.3174/ajnr.A7920
del.icio.us logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Purchase

Jump to section

  • Article
    • Abstract
    • ABBREVIATIONS:
    • MATERIALS AND METHODS
    • RESULTS
    • DISCUSSION
    • CONCLUSIONS
    • Footnotes
    • References
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF

Related Articles

  • PubMed
  • Google Scholar

Cited By...

  • AI-Generated Synthetic STIR of the Lumbar Spine from T1 and T2 MRI Sequences Trained with Open-Source Algorithms
  • The Future of Artificial Intelligence in Clinical Radiology: Savior or False Hope?
  • Crossref (11)
  • Google Scholar

This article has been cited by the following articles in journals that are participating in Crossref Cited-by Linking.

  • Diagnostic evaluation of deep learning accelerated lumbar spine MRI
    Komal M Awan, Augusto Lio M Goncalves Filho, Azadeh Tabari, Brooks P Applewhite, Min Lang, Wei-Ching Lo, Robert Sellers, Peter Kollasch, Bryan Clifford, Dominik Nickel, Jad Husseni, Otto Rapalino, Pamela Schaefer, Stephen Cauley, Susie Y Huang, John Conklin
    The Neuroradiology Journal 2024 37 3
  • Generative adversarial networks for spine imaging: A critical review of current applications
    Konstantinos Vrettos, Emmanouil Koltsakis, Aristeidis H. Zibis, Apostolos H. Karantanas, Michail E. Klontzas
    European Journal of Radiology 2024 171
  • Practical Applications of Artificial Intelligence in Spine Imaging
    Upasana Upadhyay Bharadwaj, Cynthia T. Chin, Sharmila Majumdar
    Radiologic Clinics of North America 2024 62 2
  • Applications of Artificial Intelligence and Machine Learning in Spine MRI
    Aric Lee, Wilson Ong, Andrew Makmur, Yong Han Ting, Wei Chuan Tan, Shi Wei Desmond Lim, Xi Zhen Low, Jonathan Jiong Hao Tan, Naresh Kumar, James T. P. D. Hallinan
    Bioengineering 2024 11 9
  • χ‐sepnet: Deep Neural Network for Magnetic Susceptibility Source Separation
    Minjun Kim, Sooyeon Ji, Jiye Kim, Kyeongseon Min, Hwihun Jeong, Jonghyo Youn, Taechang Kim, Jinhee Jang, Berkin Bilgic, Hyeong‐Geol Shin, Jongho Lee
    Human Brain Mapping 2025 46 2
  • AI-Generated Synthetic STIR of the Lumbar Spine from T1 and T2 MRI Sequences Trained with Open-Source Algorithms
    Alice M.L. Santilli, Mark A. Fontana, Erwin E. Xia, Zenas Igbinoba, Ek Tsoon Tan, Darryl B. Sneag, J. Levi Chazen
    American Journal of Neuroradiology 2025 46 3
  • Advancing clinical MRI exams with artificial intelligence: Japan’s contributions and future prospects
    Shohei Fujita, Yasutaka Fushimi, Rintaro Ito, Yusuke Matsui, Fuminari Tatsugami, Tomoyuki Fujioka, Daiju Ueda, Noriyuki Fujima, Kenji Hirata, Takahiro Tsuboyama, Taiki Nozaki, Masahiro Yanagawa, Koji Kamagata, Mariko Kawamura, Akira Yamada, Takeshi Nakaura, Shinji Naganawa
    Japanese Journal of Radiology 2024
  • Artificial Intelligence in Spine Imaging
    Kushal Patel, Pierce Cooper, Puneet Belani, Amish Doshi
    Magnetic Resonance Imaging Clinics of North America 2025 33 2
  • In-vivo high-resolution χ-separation at 7T
    Jiye Kim, Minjun Kim, Sooyeon Ji, Kyeongseon Min, Hwihun Jeong, Hyeong-Geol Shin, Chungseok Oh, Robert J. Fox, Ken E. Sakaie, Mark J. Lowe, Se-Hong Oh, Sina Straub, Seong-Gi Kim, Jongho Lee
    NeuroImage 2025 308
  • The Future of Artificial Intelligence in Clinical Radiology: Savior or False Hope?
    Greg Zaharchuk, Janet Mei
    American Journal of Neuroradiology 2024 45 12

More in this TOC Section

  • Bern Score Validity for SIH
  • MP2RAGE 7T in MS Lesions of the Cervical Spine
  • Resisted Inspiration for CSF-Venous Fistula
Show more Spine

Similar Articles

Advertisement

Indexed Content

  • Current Issue
  • Accepted Manuscripts
  • Article Preview
  • Past Issues
  • Editorials
  • Editor's Choice
  • Fellows' Journal Club
  • Letters to the Editor
  • Video Articles

Cases

  • Case Collection
  • Archive - Case of the Week
  • Archive - Case of the Month
  • Archive - Classic Case

Special Collections

  • AJNR Awards
  • ASNR Foundation Special Collection
  • Most Impactful AJNR Articles
  • Photon-Counting CT
  • Spinal CSF Leak Articles (Jan 2020-June 2024)

More from AJNR

  • Trainee Corner
  • Imaging Protocols
  • MRI Safety Corner

Multimedia

  • AJNR Podcasts
  • AJNR Scantastics

Resources

  • Turnaround Time
  • Submit a Manuscript
  • Submit a Video Article
  • Submit an eLetter to the Editor/Response
  • Manuscript Submission Guidelines
  • Statistical Tips
  • Fast Publishing of Accepted Manuscripts
  • Graphical Abstract Preparation
  • Imaging Protocol Submission
  • Evidence-Based Medicine Level Guide
  • Publishing Checklists
  • Author Policies
  • Become a Reviewer/Academy of Reviewers
  • News and Updates

About Us

  • About AJNR
  • Editorial Board
  • Editorial Board Alumni
  • Alerts
  • Permissions
  • Not an AJNR Subscriber? Join Now
  • Advertise with Us
  • Librarian Resources
  • Feedback
  • Terms and Conditions
  • AJNR Editorial Board Alumni

American Society of Neuroradiology

  • Not an ASNR Member? Join Now

© 2025 by the American Society of Neuroradiology All rights, including for text and data mining, AI training, and similar technologies, are reserved.
Print ISSN: 0195-6108 Online ISSN: 1936-959X

Powered by HighWire