Roy Aditi, Sural Shamik, Mukherjee Jayanta, Majumdar Arun K
School of Information Technology, Indian Institute of Technology (IIT), Kharagpur 721302, India.
IEEE Trans Inf Technol Biomed. 2008 May;12(3):366-76. doi: 10.1109/titb.2007.910352.
In this paper, we propose a hierarchical state-based model for representing an echocardiogram video. It captures the semantics of video segments from dynamic characteristics of objects present in each segment. Our objective is to provide an effective method for segmenting an echo video into view, state, and substate levels. This is motivated by the need for building efficient indexing tools to support better content management. The modeling is done using four different views, namely, short axis, long axis, apical four chamber, and apical two chamber. For view classification, an artificial neural network is trained with the histogram of a region of interest of each video frame. Object states are detected with the help of synthetic M-mode images. In contrast to traditional single M-mode, we present a novel approach named sweep M-mode for state detection. We also introduce radial M-mode for substate identification from color flow Doppler 2-D imaging. The video model described here represents the semantics of video segments using first-order predicates. Suitable operators have been defined for querying the segments. We have carried out experiments on 20 echo videos and compared the results with manual annotation done by two experts. View classification accuracy is 97.19%. Misclassification error of the state detection stage is less than 13%, which is within acceptable range since only frames at the state boundaries are found to be misclassified.
在本文中,我们提出了一种基于分层状态的模型来表示超声心动图视频。它从每个视频片段中存在的对象的动态特征捕获视频片段的语义。我们的目标是提供一种将超声视频分割为视图、状态和子状态级别的有效方法。这是由构建高效索引工具以支持更好的内容管理的需求所推动的。建模使用四个不同的视图,即短轴、长轴、心尖四腔和心尖两腔。对于视图分类,使用每个视频帧感兴趣区域的直方图训练人工神经网络。借助合成M型图像检测对象状态。与传统的单M型不同,我们提出了一种名为扫描M型的新颖方法用于状态检测。我们还引入了径向M型用于从彩色血流多普勒二维成像中识别子状态。这里描述的视频模型使用一阶谓词表示视频片段的语义。已经定义了合适的运算符来查询这些片段。我们对20个超声视频进行了实验,并将结果与两位专家的手动标注进行了比较。视图分类准确率为97.19%。状态检测阶段的误分类误差小于13%,由于仅发现状态边界处的帧被误分类,所以该误差在可接受范围内。