Joint Semantic Feature and Optical Flow Learning for Automatic Echocardiography Segmentation

Publisher:
Springer Nature
Publication Type:
Chapter
Citation:
Advanced Intelligent Computing Technology and Applications, 2024, 14868 LNCS, pp. 160-171
Issue Date:
2024-01-01
Full metadata record
The left ventricle ejection fraction is an important index for assessing cardiac function and diagnosing cardiac diseases. At present, EchoNet-Dynamic dataset is the unique large-scale resource for studying ejection fraction estimation by echocardiography. Through segmentation of the end-systolic and end-diastolic frames, the ejection fraction can be calculated based on the volumes at these phases. However, existing segmentation methods either mostly focus on single-frame segmentation and rarely consider information across consecutive frames, or they fail to effectively exploit temporal information between consecutive frames, resulting in suboptimal segmentation performance. In our study, we constructed a dual-branch spatial-temporal feature extraction model for achieving echocardiogram video segmentation. One branch was dedicated to extracting semantic features of frames under supervision, while the other branch learned the optical flows between frames in an unsupervised manner. Subsequently, we jointly trained these two branches using a temporal consistency mechanism to acquire spatial-temporal features of the frames. This approach enhances both video segmentation performance and the consistency of transition frame segmentation. Experimental results demonstrate that our proposed model achieves promising segmentation performance compared to existing methods.
Please use this identifier to cite or link to this item: