Semi-automatic motion compensation of contrast-enhanced ultrasound images from abdominal organs for perfusion analysis
Language English Country United States Media print-electronic
Document type Journal Article, Research Support, Non-U.S. Gov't
PubMed
25453358
DOI
10.1016/j.compbiomed.2014.09.014
PII: S0010-4825(14)00271-6
Knihovny.cz E-resources
- Keywords
- CEUS, Contrast-enhanced ultrasound, Motion analysis, Motion compensation, Perfusion, Perfusion modeling, Registration, Ultrasonography,
- MeSH
- Abdomen diagnostic imaging MeSH
- Crohn Disease diagnostic imaging MeSH
- Cystic Fibrosis diagnostic imaging MeSH
- Databases, Factual * MeSH
- Image Interpretation, Computer-Assisted methods MeSH
- Contrast Media administration & dosage MeSH
- Humans MeSH
- Motion MeSH
- Ultrasonography MeSH
- Check Tag
- Humans MeSH
- Male MeSH
- Female MeSH
- Publication type
- Journal Article MeSH
- Research Support, Non-U.S. Gov't MeSH
- Names of Substances
- Contrast Media MeSH
This paper presents a system for correcting motion influences in time-dependent 2D contrast-enhanced ultrasound (CEUS) images to assess tissue perfusion characteristics. The system consists of a semi-automatic frame selection method to find images with out-of-plane motion as well as a method for automatic motion compensation. Translational and non-rigid motion compensation is applied by introducing a temporal continuity assumption. A study consisting of 40 clinical datasets was conducted to compare the perfusion with simulated perfusion using pharmacokinetic modeling. Overall, the proposed approach decreased the mean average difference between the measured perfusion and the pharmacokinetic model estimation. It was non-inferior for three out of four patient cohorts to a manual approach and reduced the analysis time by 41% compared to manual processing.
Department of Biomedical Engineering Brno University of Technology Czech Republic
National Centre for Ultrasound in Gastroenterology Haukeland University Hospital Bergen Norway
References provided by Crossref.org