Dual-Energy Computed Tomography of Liver: Comparison
Please note this is a comparison between Version 1 by Masakatsu Tsurusaki and Version 3 by Conner Chen.

Dual-energy computed tomography (DECT) is an imaging technique based on data acquisition at two different energy settings. Recent advances in CT have allowed data acquisitions and simultaneous analyses of X-rays at two energy levels, and have resulted in novel developments in the field of abdominal imaging. The use of low and high X-ray tube voltages in DECT provide fused images that improve the detection of liver tumors owing to the higher contrast-to-noise ratio (CNR) of the tumor compared with the liver. The use of contrast agents in CT scanning improves image quality by enhancing the CNR and signal-to-noise ratio while reducing beam-hardening artifacts. DECT can improve detection and characterization of hepatic abnormalities, including mass lesions. The technique can also be used for the diagnosis of steatosis and iron overload.

  • computed tomography
  • dual-energy CT
  • imaging
  • hepatic steatosis
  • hepatic fibrosis
  • cirrhosis

1. Introduction

Computed tomography (CT) is one of the most important diagnostic imaging modalities for the liver. Technical advances such as faster scan times, thinner slices, multiplanar reformatting, and three-dimensional (3D) rendering have revolutionized the utility of CT. DECT is an exciting recent development that promises to increase the modality’s potential [1][2]. This improvement derives from the exploitation of the energy-dependent attenuation of X-rays at two different energies, hence, justifying the term dual energy. Conventional X-ray CT produces images based on the detection of the energy-dependent attenuation of X-rays as they pass through scanned objects. DECT extends this core concept based on the use of X-rays generated at two different energies. This resolves X-ray attenuation ambiguities and therefore allows for superior material discrimination and characterization. DECT refers to the acquisition of CT datasets at two different energy spectra. Paramount among the challenges of early DECT systems was the fact that the dual-energy data had to be obtained with two separate scans that resulted in multiple complications, such as the misregistration of the two datasets [3][4][5][6][7][8]. The technical advancements in software and hardware ion mid- to late 2000s allowed the simultaneous generation and acquisition of spectral X-rays. The recent development of dual-energy scanning increased the diagnostic value and clinical applications of CT and made DECT feasible for routine clinical use.

Computed tomography (CT) is one of the most important diagnostic imaging modalities for the liver. Technical advances such as faster scan times, thinner slices, multiplanar reformatting, and three-dimensional (3D) rendering have revolutionized the utility of CT. DECT is an exciting recent development that promises to increase the modality’s potential [1,2]. This improvement derives from the exploitation of the energy-dependent attenuation of X-rays at two different energies, hence, justifying the term dual energy. Conventional X-ray CT produces images based on the detection of the energy-dependent attenuation of X-rays as they pass through scanned objects. DECT extends this core concept based on the use of X-rays generated at two different energies. This resolves X-ray attenuation ambiguities and therefore allows for superior material discrimination and characterization. DECT refers to the acquisition of CT datasets at two different energy spectra. Paramount among the challenges of early DECT systems was the fact that the dual-energy data had to be obtained with two separate scans that resulted in multiple complications, such as the misregistration of the two datasets [3,4,5,6,7,8]. The technical advancements in software and hardware ion mid- to late 2000s allowed the simultaneous generation and acquisition of spectral X-rays. The recent development of dual-energy scanning increased the diagnostic value and clinical applications of CT and made DECT feasible for routine clinical use.

2. Evaluation of Hepatic Steatosis Using DECT

Conventionally available methods include MR spectroscopy (MRS), the dual-echo MRI method, CT measurements, and visual evaluation via ultrasonic tests. Among these, MRS (magnetic resonance spectroscopy) has yielded the highest fat content [9]. Conversely, new quantitative methods have been developed for ultrasound tests, MRI, and CT. According to a meta-analysis [10], the detectability of moderate/severe (20–30%) fatty liver using B-mode ultrasound-based qualitative assessment had a sensitivity of 84.8% (95% CI: 79.5 to 88.9) and a specificity of 93.6% (87.2 to 97.0). Chemical shift imaging is said to improve quantitativeness by correcting the influence of T1 value, T2* attenuation, variation in precession frequency of fat protons, etc. [11][12]. A few studies have measured the amount of fat obtained through T1-independent, T2*-corrected chemical shift imaging with multi-peak fat spectral modeling relative to the MRS-proton density fat fraction (MRS-PDFF) measured using MRS, as MRI-PDFF [12]. Although it is recognized that the MR test is the most accurate for fat determination among each of the imaging modalities, there are variations in the method and it is difficult to construct coherent evidence at the present time [13]. As fibrosis occurs due to hepatic injury, the amount of fat decreases; thus, low fat quantity detected on imaging does not guarantee a milder case and this is applicable for not only MRI, but also CT and ultrasound tests. In CT, the semiquantitative evaluation using X-ray absorption values (HUs) has been extensively known. Although it is convenient to conduct these evaluations based on the CT value of the liver parenchyma, the CT value may not necessarily be consistent among various CT scanners. Therefore, the evaluation with the use of a ratio or difference in CT values between liver and spleen should be preferred. Park et al. [14] argued that the difference in liver to spleen CT values in conventional CT had a reference range between 1–18 HU in nonfatty livers, where 1 HU (or less) is indicative of a fatty liver. A few recent reports have suggested that with DECT imaging—that is becoming widespread in recent years—it is possible to quantify the amount of liver fat in contrast CT based on material decomposition that helps distinguish two or more substances [15][16][17][18][19] (

Conventionally available methods include MR spectroscopy (MRS), the dual-echo MRI method, CT measurements, and visual evaluation via ultrasonic tests. Among these, MRS (magnetic resonance spectroscopy) has yielded the highest fat content [35]. Conversely, new quantitative methods have been developed for ultrasound tests, MRI, and CT. According to a meta-analysis [36], the detectability of moderate/severe (20–30%) fatty liver using B-mode ultrasound-based qualitative assessment had a sensitivity of 84.8% (95% CI: 79.5 to 88.9) and a specificity of 93.6% (87.2 to 97.0). Chemical shift imaging is said to improve quantitativeness by correcting the influence of T1 value, T2* attenuation, variation in precession frequency of fat protons, etc. [37,38]. A few studies have measured the amount of fat obtained through T1-independent, T2*-corrected chemical shift imaging with multi-peak fat spectral modeling relative to the MRS-proton density fat fraction (MRS-PDFF) measured using MRS, as MRI-PDFF [38]. Although it is recognized that the MR test is the most accurate for fat determination among each of the imaging modalities, there are variations in the method and it is difficult to construct coherent evidence at the present time [39]. As fibrosis occurs due to hepatic injury, the amount of fat decreases; thus, low fat quantity detected on imaging does not guarantee a milder case and this is applicable for not only MRI, but also CT and ultrasound tests. In CT, the semiquantitative evaluation using X-ray absorption values (HUs) has been extensively known. Although it is convenient to conduct these evaluations based on the CT value of the liver parenchyma, the CT value may not necessarily be consistent among various CT scanners. Therefore, the evaluation with the use of a ratio or difference in CT values between liver and spleen should be preferred. Park et al. [40] argued that the difference in liver to spleen CT values in conventional CT had a reference range between 1–18 HU in nonfatty livers, where 1 HU (or less) is indicative of a fatty liver. A few recent reports have suggested that with DECT imaging—that is becoming widespread in recent years—it is possible to quantify the amount of liver fat in contrast CT based on material decomposition that helps distinguish two or more substances [41,42,43,44,45] (

Figure 6

).

Figure 6.

 Liver fat measurements by using dual-energy CT in the patient with NAFLD (non-alcoholic fatty liver disease)(

a

) Noncontrast CT image and (

b

) fat density image in patients with nonalcoholic fatty liver disease. The degree of fat deposition was quantifiable on the fat density images, and the quantitative assessment of liver tissue and fat in this case was 22% liver-fat quantification.

Hyodo et al. reported that the area under the receiver operating characteristic (ROC) curve of DECT was 0.88 (95% confidence interval 0.74–0.98) in the detection of fatty liver with more than 5% fat in needle biopsy tissue. This was equivalent to MRS (area under the ROC curve 0.89) (95% confidence interval 0.724–1) [15]. This study also suggested that high-precision fat quantification is possible in postcontrast images. Hyodo et al. also reported that fat was underestimated by the presence of iron deposited in the liver parenchyma [16], while the phantom study conducted by Fishcer et al. showed that accurate fat quantitation was possible even in the presence of iron [17].

Hyodo et al. reported that the area under the receiver operating characteristic (ROC) curve of DECT was 0.88 (95% confidence interval 0.74–0.98) in the detection of fatty liver with more than 5% fat in needle biopsy tissue. This was equivalent to MRS (area under the ROC curve 0.89) (95% confidence interval 0.724–1) [41]. This study also suggested that high-precision fat quantification is possible in postcontrast images. Hyodo et al. also reported that fat was underestimated by the presence of iron deposited in the liver parenchyma [42], while the phantom study conducted by Fishcer et al. showed that accurate fat quantitation was possible even in the presence of iron [43].

Evaluation of active inflammation in addition to fat is crucial for the early diagnosis and staging of NASH. Because it is difficult to evaluate inflammation with imaging studies with CT alone, the future challenge will involve accurate assessments in conjunction with other examinations.

3. Estimation of Hepatic Fibrosis Using DECT

Given the degree of hepatic fibrosis correlates with carcinogenesis and prognosis, accurate evaluation is important in determining the clinical course and in predicting the prognosis. Non-invasive diagnosis and staging of hepatic fibrosis are important for evaluating disease progression in patients with chronic liver diseases. Serum markers of liver fibrosis are widely available, but their results are variable. Ultrasound elastography is used for grading hepatic fibrosis, but it is operator dependent [20][21][22]. Several magnetic resonance (MR) imaging sequences, such as perfusion-weighted MR imaging, MR spectroscopy, MR elastography, and dynamic contrast-enhanced MR have the potential to provide quantitative information with variable diagnostic accuracy in the staging of liver fibrosis. DWI provides noninvasive quantification of water diffusion and can be used for in vivo quantification of the combined effects of capillary perfusion and diffusion. There are limited data on correlation between ADC and degree of fibrosis and other histopathological findings (inflammation, steatosis, iron content, necrosis, cholestasis) in the liver [23]. Besheer et al. reported that the ADC value was decreased from controls (F0) to patients with early fibrosis and those with late fibrosis. Combined ADC and miR-200b revealed the best result for differentiating early from late fibrosis and offer an alternative surrogate non-invasive diagnostic tool for diagnosis and staging of hepatic fibrosis in patients with chronic hepatitis C [24].

Given the degree of hepatic fibrosis correlates with carcinogenesis and prognosis, accurate evaluation is important in determining the clinical course and in predicting the prognosis. Non-invasive diagnosis and staging of hepatic fibrosis are important for evaluating disease progression in patients with chronic liver diseases. Serum markers of liver fibrosis are widely available, but their results are variable. Ultrasound elastography is used for grading hepatic fibrosis, but it is operator dependent [46,47,48]. Several magnetic resonance (MR) imaging sequences, such as perfusion-weighted MR imaging, MR spectroscopy, MR elastography, and dynamic contrast-enhanced MR have the potential to provide quantitative information with variable diagnostic accuracy in the staging of liver fibrosis. DWI provides noninvasive quantification of water diffusion and can be used for in vivo quantification of the combined effects of capillary perfusion and diffusion. There are limited data on correlation between ADC and degree of fibrosis and other histopathological findings (inflammation, steatosis, iron content, necrosis, cholestasis) in the liver [49]. Besheer et al. reported that the ADC value was decreased from controls (F0) to patients with early fibrosis and those with late fibrosis. Combined ADC and miR-200b revealed the best result for differentiating early from late fibrosis and offer an alternative surrogate non-invasive diagnostic tool for diagnosis and staging of hepatic fibrosis in patients with chronic hepatitis C [50].

There are insufficient differences in the mass attenuation coefficients between fibrous and normal liver tissue, and it is difficult to quantify hepatic fibrosis by SECT. However, it is possible to indirectly estimate the degree of fibrosis with contrast media as a marker. When fibrosis occurs in liver tissue, deposition of collagen fibers expands the extracellular space. Histologically, the degree of hepatic fibrosis is strongly correlated with the volume of the extracellular space (ECV), and the quantification of ECV can be used to estimate the degree of hepatic fibrosis [25]. The equilibrium phase of contrast-enhanced CT is used for the determination of ECV in liver tissue. Given that the administered contrast medium diffuses from the intravascular space into the extravascular space and equilibrates over time, the concentration of contrast medium in the intravascular and extravascular spaces is considered to be equal during the equilibrium phase of contrast-enhanced CT. Therefore, by normalizing the contrast enhancement in the liver parenchyma in the equilibrium phase by the contrast enhancement in the blood pool, such as in the aorta, the percentage of ECV fraction in the liver tissue (ECV fraction) can be measured, and the ECV fraction can thus be calculated using the formula: (enhancement in the liver)/(enhancement in the aorta) × (1 − hematocrit).

There are insufficient differences in the mass attenuation coefficients between fibrous and normal liver tissue, and it is difficult to quantify hepatic fibrosis by SECT. However, it is possible to indirectly estimate the degree of fibrosis with contrast media as a marker. When fibrosis occurs in liver tissue, deposition of collagen fibers expands the extracellular space. Histologically, the degree of hepatic fibrosis is strongly correlated with the volume of the extracellular space (ECV), and the quantification of ECV can be used to estimate the degree of hepatic fibrosis [51]. The equilibrium phase of contrast-enhanced CT is used for the determination of ECV in liver tissue. Given that the administered contrast medium diffuses from the intravascular space into the extravascular space and equilibrates over time, the concentration of contrast medium in the intravascular and extravascular spaces is considered to be equal during the equilibrium phase of contrast-enhanced CT. Therefore, by normalizing the contrast enhancement in the liver parenchyma in the equilibrium phase by the contrast enhancement in the blood pool, such as in the aorta, the percentage of ECV fraction in the liver tissue (ECV fraction) can be measured, and the ECV fraction can thus be calculated using the formula: (enhancement in the liver)/(enhancement in the aorta) × (1 − hematocrit).

In SECT, the difference between pre- and postcontrast CT values is used to quantify the ECV fraction or to estimate hepatic fibrosis. The time from the injection of contrast medium to the imaging of the equilibrium phase varied between 3 and 30 min [26][27][28][29]. Yoon et al. and Guo et al. reported that the ECV fraction calculated using the equilibrium phase 3 min after contrast was significantly correlated with the METAVIR score (correlation coefficient 

In SECT, the difference between pre- and postcontrast CT values is used to quantify the ECV fraction or to estimate hepatic fibrosis. The time from the injection of contrast medium to the imaging of the equilibrium phase varied between 3 and 30 min [52,53,54,55]. Yoon et al. and Guo et al. reported that the ECV fraction calculated using the equilibrium phase 3 min after contrast was significantly correlated with the METAVIR score (correlation coefficient 

ρ

 = 0.493 and 

ρ = 0.468, respectively) [28][29]. In addition, the ECV fraction calculated using nonlinear nonrigid alignment (SURE subtraction) from the equilibrium phase 4 min after contrast enhancement, and the pre-contrast images, showed a strong correlation with the METAVIR score (

 = 0.468, respectively) [54,55]. In addition, the ECV fraction calculated using nonlinear nonrigid alignment (SURE subtraction) from the equilibrium phase 4 min after contrast enhancement, and the pre-contrast images, showed a strong correlation with the METAVIR score (

ρ = 0.71) [30]. In DECT, a more quantitative ECV fraction can be estimated based on the use of the difference in iodine density between pre- and postcontrast images. Sofue et al. reported that the ECV fraction measured using iodine density images in the equilibrium phase 3 min after contrast was significantly correlated with the METAVIR score (

 = 0.71) [56]. In DECT, a more quantitative ECV fraction can be estimated based on the use of the difference in iodine density between pre- and postcontrast images. Sofue et al. reported that the ECV fraction measured using iodine density images in the equilibrium phase 3 min after contrast was significantly correlated with the METAVIR score (

ρ = 0.67) [31]. The effect of hepatic active inflammation and edema on the quantitation of ECV is also unclear. These verifications and further improvements in the quantitative accuracy seem to be future problems.

 = 0.67) [57]. The effect of hepatic active inflammation and edema on the quantitation of ECV is also unclear. These verifications and further improvements in the quantitative accuracy seem to be future problems.