共查询到20条相似文献,搜索用时 671 毫秒
1.
2.
4.
5.
多目标多传感器的跟踪是一个非常复杂的问题,其关键是数据的融合。单个传感器的多目标跟踪主要是解决数据的关联问题,多个传感器的多目标跟踪虽然有时也是解决数据关联问题,但是多传感器的数据关联就要复杂得多,它是一个多维的问题。文中主要介绍多目标多传感器跟踪网络构成策略及分级分布传感器跟踪的数据和轨迹融合方式。 相似文献
6.
本文叙述微机应用中一个重要领域——多总线和多微机处理系统,包括它们的结构、类型定义、主要连接模块、信息传送和总线软件特点等。可供设计人员参考和借鉴。 相似文献
7.
所谓定时控制器就是按照预定的程序控制执行机构。完成一系列操作。它可广泛应用于工业自动化生产的某些操作过程。如化工厂反应罐的升温、降温、添料、冷却;电镀车间的镀件、取件、清洗等;也可用于学校和企事业单位作息的自动打铃以及各种家用电器的定时启闭等。根据不同的原理,采用不同的电路,选用不同的元器件,均可实现上述功能。本文介绍的时控器采用大规模集成电路EPROM完成译码和动态扫描显示,用CMOS RAM作为操作内容的存贮单元,故系统具有电路简单,操作方便,抗干扰能力强等特点,并真正做到了“多路”、“多点”。一、功能特点 相似文献
8.
9.
10.
多尺度多特征仿生人脸识别 总被引:1,自引:1,他引:0
本文使用Daubechies正交小波变换对人脸图像进行二次小波分解:首先对第二次小波变换低频子图像进行PCA分析。运用邻域法进行分类得到距离隶属度。利用模糊分析提取出候选样本,对候选样本第一次小波变换的低频子图像进行PCA分析,运用最近邻域法进行分类得到最终识别结果。实验表明:小波变换预处理得到多尺度多特征;分类结果之间具有一定的互补性,同时可以提高分类性能。 相似文献
11.
Hyun-seok Min Jae Young Choi Wesley De Neve Yong Man Ro 《Signal Processing: Image Communication》2011,26(10):612-627
The detection of near-duplicate video clips (NDVCs) is an area of current research interest and intense development. Most NDVC detection methods represent video clips with a unique set of low-level visual features, typically describing color or texture information. However, low-level visual features are sensitive to transformations of the video content. Given the observation that transformations tend to preserve the semantic information conveyed by the video content, we propose a novel approach for identifying NDVCs, making use of both low-level visual features (this is, MPEG-7 visual features) and high-level semantic features (this is, 32 semantic concepts detected using trained classifiers). Experimental results obtained for the publicly available MUSCLE-VCD-2007 and TRECVID 2008 video sets show that bimodal fusion of visual and semantic features facilitates robust NDVC detection. In particular, the proposed method is able to identify NDVCs with a low missed detection rate (3% on average) and a low false alarm rate (2% on average). In addition, the combined use of visual and semantic features outperforms the separate use of either of them in terms of NDVC detection effectiveness. Further, we demonstrate that the effectiveness of the proposed method is on par with or better than the effectiveness of three state-of-the-art NDVC detection methods either making use of temporal ordinal measurement, features computed using the Scale-Invariant Feature Transform (SIFT), or bag-of-visual-words (BoVW). We also show that the influence of the effectiveness of semantic concept detection on the effectiveness of NDVC detection is limited, as long as the mean average precision (MAP) of the semantic concept detectors used is higher than 0.3. Finally, we illustrate that the computational complexity of our NDVC detection method is competitive with the computational complexity of the three aforementioned NDVC detection methods. 相似文献
12.
Andre B Vercauteren T Buchner AM Wallace MB Ayache N 《IEEE transactions on medical imaging》2012,31(6):1276-1288
Content-based image retrieval (CBIR) is a valuable computer vision technique which is increasingly being applied in the medical community for diagnosis support. However, traditional CBIR systems only deliver visual outputs, i.e., images having a similar appearance to the query, which is not directly interpretable by the physicians. Our objective is to provide a system for endomicroscopy video retrieval which delivers both visual and semantic outputs that are consistent with each other. In a previous study, we developed an adapted bag-of-visual-words method for endomicroscopy retrieval, called "Dense-Sift," that computes a visual signature for each video. In this paper, we present a novel approach to complement visual similarity learning with semantic knowledge extraction, in the field of in vivo endomicroscopy. We first leverage a semantic ground truth based on eight binary concepts, in order to transform these visual signatures into semantic signatures that reflect how much the presence of each semantic concept is expressed by the visual words describing the videos. Using cross-validation, we demonstrate that, in terms of semantic detection, our intuitive Fisher-based method transforming visual-word histograms into semantic estimations outperforms support vector machine (SVM) methods with statistical significance. In a second step, we propose to improve retrieval relevance by learning an adjusted similarity distance from a perceived similarity ground truth. As a result, our distance learning method allows to statistically improve the correlation with the perceived similarity. We also demonstrate that, in terms of perceived similarity, the recall performance of the semantic signatures is close to that of visual signatures and significantly better than those of several state-of-the-art CBIR methods. The semantic signatures are thus able to communicate high-level medical knowledge while being consistent with the low-level visual signatures and much shorter than them. In our resulting retrieval system, we decide to use visual signatures for perceived similarity learning and retrieval, and semantic signatures for the output of an additional information, expressed in the endoscopist own language, which provides a relevant semantic translation of the visual retrieval outputs. 相似文献
13.
This paper presents a new learning algorithm for audiovisual fusion and demonstrates its application to video classification
for film database. The proposed system utilized perceptual features for content characterization of movie clips. These features
are extracted from different modalities and fused through a machine learning process. More specifically, in order to capture
the spatio-temporal information, an adaptive video indexing is adopted to extract visual feature, and the statistical model
based on Laplacian mixture are utilized to extract audio feature. These features are fused at the late fusion stage and input
to a support vector machine (SVM) to learn semantic concepts from a given video database. Based on our experimental results,
the proposed system implementing the SVM-based fusion technique achieves high classification accuracy when applied to a large
volume database containing Hollywood movies. 相似文献
14.
A. G. Hauptmann M.-Y. Chen M. Christel W.-H. Lin J. Yang 《Journal of Signal Processing Systems》2010,58(3):373-385
In this paper we describe a multi-strategy approach to improving semantic extraction from news video. Experiments show the
value of careful parameter tuning, exploiting multiple feature sets and multilingual linguistic resources, applying text retrieval
approaches for image features, and establishing synergy between multiple concepts through undirected graphical models. We
present a discriminative learning framework called Multi-concept Discriminative Random Field (MDRF) for building probabilistic
models of video semantic concept detectors by incorporating related concepts as well as the low-level observations. The model
exploits the power of discriminative graphical models to simultaneously capture the associations of concept with observed
data and the interactions between related concepts. Compared with previous methods, this model not only captures the co-occurrence
between concepts but also incorporates the raw data observations into a unified framework. We also describe an approximate
parameter estimation algorithm and present results obtained from the TRECVID 2006 data. No single approach, however, provides
a consistently better result for all concept detection tasks, which suggests that extracting video semantics should exploit
multiple resources and techniques rather than naively relying on a single approach 相似文献
15.
Video semantic detection has been one research hotspot in the field of human-computer interaction. In video features-oriented sparse representation, the features from the same category video could not achieve similar coding results. To address this, the Locality-Sensitive Discriminant Sparse Representation (LSDSR) is developed, in order that the video samples belonging to the same video category are encoded as similar sparse codes which make them have better category discrimination. In the LSDSR, a discriminative loss function based on sparse coefficients is imposed on the locality-sensitive sparse representation, which makes the optimized dictionary for sparse representation be discriminative. The LSDSR for video features enhances the power of semantic discrimination to optimize the dictionary and build the better discriminant sparse model. More so, to further improve the accuracy of video semantic detection after sparse representation, a weighted K-Nearest Neighbor (KNN) classification method with the loss function that integrates reconstruction error and discrimination for the sparse representation is adopted to detect video semantic concepts. The proposed methods are evaluated on the related video databases in comparison with existing sparse representation methods. The experimental results show that the proposed methods significantly enhance the power of discrimination of video features, and consequently improve the accuracy of video semantic concept detection. 相似文献
16.
Jiang YG Dai Q Wang J Ngo CW Xue X Chang SF 《IEEE transactions on image processing》2012,21(6):3080-3091
Exploring context information for visual recognition has recently received significant research attention. This paper proposes a novel and highly efficient approach, which is named semantic diffusion, to utilize semantic context for large-scale image and video annotation. Starting from the initial annotation of a large number of semantic concepts (categories), obtained by either machine learning or manual tagging, the proposed approach refines the results using a graph diffusion technique, which recovers the consistency and smoothness of the annotations over a semantic graph. Different from the existing graph-based learning methods that model relations among data samples, the semantic graph captures context by treating the concepts as nodes and the concept affinities as the weights of edges. In particular, our approach is capable of simultaneously improving annotation accuracy and adapting the concept affinities to new test data. The adaptation provides a means to handle domain change between training and test data, which often occurs in practice. Extensive experiments are conducted to improve concept annotation results using Flickr images and TV program videos. Results show consistent and significant performance gain (10 +% on both image and video data sets). Source codes of the proposed algorithms are available online. 相似文献
17.
18.
19.
20.
The analysis of moving objects in videos, especially the recognition of human motions and gestures, is attracting increasing emphasis in computer vision area. However, most existing video analysis methods do not take into account the effect of video semantic information. The topological information of the video image plays an important role in describing the association relationship of the image content, which will help to improve the discriminability of the video feature expression. Based on the above considerations, we propose a video semantic feature learning method that integrates image topological sparse coding with dynamic time warping algorithm to improve the gesture recognition in videos. This method divides video feature learning into two phases: semi-supervised video image feature learning and supervised optimization of video sequence features. Next, a distance weighting based dynamic time warping algorithm and K-nearest neighbor algorithm is leveraged to recognize gestures. We conduct comparative experiments on table tennis video dataset. The experimental results show that the proposed method is more discriminative to the expression of video features and can effectively improve the recognition rate of gestures in sports video. 相似文献