Browsing by Subject "Video recording"
Now showing 1 - 16 of 16
- Results Per Page
- Sort Options
Item Open Access 2-D triangular mesh-based mosaicking for object tracking in the presence of occlusion(SPIE, 1997) Toklu, C.; Tekalp, A. M.; Erdem, A. TanjuIn this paper, we describe a method for temporal tracking of video objects in video clips. We employ a 2D triangular mesh to represent each video object, which allows us to describe the motion of the object by the displacements of the node points of the mesh, and to describe any intensity variations by the contrast and brightness parameters estimated for each node point. Using the temporal history of the node point locations, we continue tracking the nodes of the 2D mesh even when they become invisible because of self-occlusion or occlusion by another object. Uncovered parts of the object in the subsequent frames of the sequence are detected by means of an active contour which contains a novel shape preserving energy term. The proposed shape preserving energy term is found to be successful in tracking the boundary of an object in video sequences with complex backgrounds. By adding new nodes or updating the 2D triangular mesh we incrementally append the uncovered parts of the object detected during the tracking process to the one of the objects to generate a static mosaic of the object. Also, by texture mapping the covered pixels into the current frame of the video clip we can generate a dynamic mosaic of the object. The proposed mosaicing technique is more general than those reported in the literature because it allows for local motion and out-of-plane rotations of the object that results in self-occlusions. Experimental results demonstrate the successful tracking of the objects with deformable boundaries in the presence of occlusion.Item Open Access 3D human pose search using oriented cylinders(IEEE, 2009-09-10) Pehlivan, Selen; Duygulu, PınarIn this study, we present a representation based on a new 3D search technique for volumetric human poses which is then used to recognize actions in three dimensional video sequences. We generate a set of cylinder like 3D kernels in various sizes and orientations. These kernels are searched over 3D volumes to find high response regions. The distribution of these responses are then used to represent a 3D pose. We use the proposed representation for (i) pose retrieval using Nearest Neighbor (NN) based classification and Support Vector Machine (SVM) based classification methods, and for (ii) action recognition on a set of actions using Dynamic Time Warping (DTW) and Hidden Markov Model (HMM) based classification methods. Evaluations on IXMAS dataset supports the effectiveness of such a robust pose representation. ©2009 IEEE.Item Open Access Automatic detection of salient objects and spatial relations in videos for a video database system(Elsevier BV, 2008-10) Sevilmiş, T.; Baştan M.; Güdükbay, Uğur; Ulusoy, ÖzgürMultimedia databases have gained popularity due to rapidly growing quantities of multimedia data and the need to perform efficient indexing, retrieval and analysis of this data. One downside of multimedia databases is the necessity to process the data for feature extraction and labeling prior to storage and querying. Huge amount of data makes it impossible to complete this task manually. We propose a tool for the automatic detection and tracking of salient objects, and derivation of spatio-temporal relations between them in video. Our system aims to reduce the work for manual selection and labeling of objects significantly by detecting and tracking the salient objects, and hence, requiring to enter the label for each object only once within each shot instead of specifying the labels for each object in every frame they appear. This is also required as a first step in a fully-automatic video database management system in which the labeling should also be done automatically. The proposed framework covers a scalable architecture for video processing and stages of shot boundary detection, salient object detection and tracking, and knowledge-base construction for effective spatio-temporal object querying. © 2008 Elsevier B.V. All rights reserved.Item Open Access Cepstrum based method for moving shadow detection in video(Springer, 2010-09) Cogun, Fuat; Çetin, A. EnisMoving shadows constitute problems in various applications such as image segmentation and object tracking. Main cause of these problems is the misclassification of the shadow pixels as target pixels. Therefore, the use of an accurate and reliable shadow detection method is essential to realize intelligent video processing applications. In this paper, the cepstrum based method for moving shadow detection is presented. The proposed method is tested on outdoor and indoor video sequences using well-known benchmark test sets. To show the improvements over previous approaches, quantitative metrics are introduced and comparisons based on these metrics are made. © 2011 Springer Science+Business Media B.V.Item Open Access Computer vision based text and equation editor for LATEX(IEEE, 2004-06) Öksüz, Özcan; Güdükbay, Uğur; Çetin, EnisIn this paper, we present a computer vision based text and equation editor for LATEX. The user writes text and equations on paper and a camera attached to a computer records actions of the user. In particular, positions of the pen-tip in consecutive image frames are detected. Next, directional and positional information about characters are calculated using these positions. Then, this information is used for on-line character classification. After characters and symbols are found, corresponding LATEX code is generated.Item Open Access Flame detection in video using hidden Markov models(IEEE, 2005) Töreyin, B. Uğur; Dedeoğlu, Yiğithan; Çetin, A. EnisThis paper proposes a novel method to detect flames in video by processing the data generated by an ordinary camera monitoring a scene. In addition to ordinary motion and color clues, flame flicker process is also detected by using a hidden Markov model. Markov models representing the flame and flame colored ordinary moving objects are used to distinguish flame flicker process from motion of flame colored moving objects. Spatial color variations in flame are also evaluated by the same Markov models, as well. These clues are combined to reach a final decision. False alarms due to ordinary motion of flame colored moving objects are greatly reduced when compared to the existing video based fire detection systems.Item Open Access Flame detection method in video using covariance descriptors(IEEE, 2011) Habiboǧlu, Y.H.; Günay, Osman; Çetin, A. EnisVideo fire detection system which uses a spatio-temporal covariance matrix of video data is proposed. This system divides the video into spatio-temporal blocks and computes covariance features extracted from these blocks to detect fire. Feature vectors taking advantage of both the spatial and the temporal characteristics of flame colored regions are classified using an SVM classifier which is trained and tested using video data containing flames and flame colored objects. Experimental results are presented. © 2011 IEEE.Item Open Access HandVR: a hand-gesture-based interface to a video retrieval system(Springer U K, 2015) Genç, S.; Baştan M.; Güdükbay, Uğur; Atalay, V.; Ulusoy, ÖzgürUsing one’s hands in human–computer interaction increases both the effectiveness of computer usage and the speed of interaction. One way of accomplishing this goal is to utilize computer vision techniques to develop hand-gesture-based interfaces. A video database system is one application where a hand-gesture-based interface is useful, because it provides a way to specify certain queries more easily. We present a hand-gesture-based interface for a video database system to specify motion and spatiotemporal object queries. We use a regular, low-cost camera to monitor the movements and configurations of the user’s hands and translate them to video queries. We conducted a user study to compare our gesture-based interface with a mouse-based interface on various types of video queries. The users evaluated the two interfaces in terms of different usability parameters, including the ease of learning, ease of use, ease of remembering (memory), naturalness, comfortable use, satisfaction, and enjoyment. The user study showed that querying video databases is a promising application area for hand-gesture-based interfaces, especially for queries involving motion and spatiotemporal relations.Item Open Access Key frame selection from MPEG video data(SPIE, 1997-02) Gerek, Ömer. N.; Altunbaşak, Y.This paper describes a method for selecting key frames by using a number of parameters extracted from the MPEG video stream. The parameters are directly extracted from the compressed video stream without decompression. A combination of these parameters are then used in a rule based decision system. The computational complexity for extracting the parameters and for key frame decision rule is very small. As a results, the overall operation is very quickly performed and this makes our algorithm handy for practical purposes. The experimental results show that this method can select the distinctive frames of video streams successfully.Item Open Access MPEG-7 uyumlu video veri tabanlari için önemli nesnelerin otomatik olarak bulunmasi(IEEE, 2008-04) Baştan, Muhammed; Güdükbay, Uğur; Ulusoy, ÖzgürBu çalışma, genel olarak nesneye dayalı endekslemeyi destekleyen, özel olarak MPEG-7 uyumlu veritabanları için, videolardan önemli nesnelerin otomatik olarak çıkarılmasını saglayabilecek bir yöntem sunmaktadır. Şimdiye kadar yapılan benzer çalışmalar genellikle resimler üzerinde yoğunlaşmış ve sadece ilk bakışta dikkati çeken alanları bulmaya çalışmıştır. Önerilen yöntem ise videolar üzerinde çalışmak için tasarlanmış olup sadece ilk bakışta dikkat çeken bölgelerin değil, videonun endekslenmesi için önemli sayılabilecek bölgelerin de bulunabilmesini amaçlamaktadır. Bunun için önce video kareleri bölütlere ayrılmakta, sonra her bölüt için yerel ve genel renk, biçim, doku ve hareket bilgileri hesaplanmakta, son olarak bu özellikler kullanılarak eğitilmiş bir destek vektor makinesi (SVM) kullanılarak bölgelerin önemli olup olmadığına karar verilmektedir. İlk deney sonuçları önerilen y öntemin başarılı olduğunu ve elde edilen nesnelerin öncekilere g öre anlamsal olarak daha iyi olduğunu göstermektedir. We describe a method to automatically extract video objects, which are important for object-based indexing of videos in an MPEG-7 compliant video database system. Most of the existing salient object detection approaches detect visually conspicuous image structures, while our method aims to find regions that may be important for indexing in a video database system. Our method works on a shot basis. We first segment each frame to obtain homogeneous regions in terms of color and texture. Then, we extract a set of local and global color, shape, texture and motion features for each region. Finally, the regions are classified as being salient or non-salient using SVMs trained on a few hundreds of example regions. Experimental results from news video segments show that the proposed method is more effective in extracting the important regions in terms of human visual perception. ©2008 IEEE.Item Open Access Scalable image quality assessment with 2D mel-cepstrum and machine learning approach(Elsevier, 2011-07-19) Narwaria, M.; Lin, W.; Çetin, A. EnisMeasurement of image quality is of fundamental importance to numerous image and video processing applications. Objective image quality assessment (IQA) is a two-stage process comprising of the following: (a) extraction of important information and discarding the redundant one, (b) pooling the detected features using appropriate weights. These two stages are not easy to tackle due to the complex nature of the human visual system (HVS). In this paper, we first investigate image features based on two-dimensional (2D) mel-cepstrum for the purpose of IQA. It is shown that these features are effective since they can represent the structural information, which is crucial for IQA. Moreover, they are also beneficial in a reduced-reference scenario where only partial reference image information is used for quality assessment. We address the second issue by exploiting machine learning. In our opinion, the well established methodology of machine learning/pattern recognition has not been adequately used for IQA so far; we believe that it will be an effective tool for feature pooling since the required weights/parameters can be determined in a more convincing way via training with the ground truth obtained according to subjective scores. This helps to overcome the limitations of the existing pooling methods, which tend to be over simplistic and lack theoretical justification. Therefore, we propose a new metric by formulating IQA as a pattern recognition problem. Extensive experiments conducted using six publicly available image databases (totally 3211 images with diverse distortions) and one video database (with 78 video sequences) demonstrate the effectiveness and efficiency of the proposed metric, in comparison with seven relevant existing metrics.Item Open Access Segmentation-based extraction of important objects from video for object-based indexing(IEEE, 2008-06) Baştan, Muhammet; Güdükbay, Uğur; Ulusoy, ÖzgürWe describe a method to automatically extract important video objects for object-based indexing. Most of the existing salient object detection approaches detect visually conspicuous structures in images, while our method aims to find regions that may be important for indexing in a video database system. Our method works on a shot basis. We first segment each frame to obtain homogeneous regions in terms of color and texture. Then, we extract a set of regional and inter-regional color, shape, texture and motion features for all regions, which are classified as being important or not using SVMs trained on a few hundreds of example regions. Finally, each important region is tracked within each shot for trajectory generation and consistency check. Experimental results from news video sequences show that the proposed approach is effective. © 2008 IEEE.Item Open Access Ses ve video işaretlerinde saklı markof modeli tabanlı düşen kişi tespiti(IEEE, 2006-04) Töreyin, B. Uğur; Dedeoğlu, Yiğithan; Çetin, A. EnisAutomatic detection of a falling person in video is an important problem with applications in security and safety areas including supportive home environments and CCTV surveillance systems. Human motion in video is modeled using Hidden Markov Models (HMM) in this paper. In addition, the audio track of the video is also used to distinguish a person simply sitting on a floor from a person stumbling and falling. Most video recording systems have the capability of recording audio as well and the impact sound of a falling person is also available as an additional clue. Audio channel data based decision is also reached using HMMs and fused with results of HMMs modeling the video data to reach a final decision. © 2006 IEEE.Item Open Access Summarization of documentaries(Springer, Dordrecht, 2010) Demirtas, K.; Çiçekli, İlyas; Cicekli, N.K.Video summarization algorithms present condensed versions of a full length video by identifying the most significant parts of the video. In this paper, we propose an automatic video summarization method using the subtitles of videos and text summarization techniques. We identify significant sentences in the subtitles of a video by using text summarization techniques and then we compose a video summary by finding the video parts corresponding to these summary sentences. © 2011 Springer Science+Business Media B.V.Item Open Access Universal product design involving elderly users: a participatory design model(Elsevier, 2004-07) Demirbilek, O.; Demirkan, H.Recent studies have shown that people prefer to age in their familiar environments, thus guiding designers to provide a safe and functionally appropriate environment for ageing people, regardless of their physical conditions or limitations. Therefore, a participatory design model is proposed where human beings can improve their quality of life by promoting independence, as well as safety, useability and attractiveness of the residence. Brainstorming, scenario building, unstructured interviews, sketching and videotaping are used as techniques in the participatory design sessions. Quality deployment matrices are employed to find the relationships between the elderly user's requirements and design specifications. A case study was devised to apply and test the conceptual model phase of the proposed model.Item Open Access Wildfire detection using LMS based active learning(IEEE, 2009-04) Töreyin, B. Uğur; Çetin, A. EnisA computer vision based algorithm for wildfire detection is developed. The main detection algorithm is composed of four sub-algorithms detecting (i) slow moving objects, (ii) gray regions, (iii) rising regions, and (iv) shadows. Each algorithm yields its own decision as a real number in the range [-1,1] at every image frame of a video sequence. Decisions from subalgorithms are fused using an adaptive algorithm. In contrast to standard Weighted Majority Algorithm (WMA), weights are updated using the Least Mean Square (LMS) method in the training (learning) stage. The error function is defined as the difference between the overall decision of the main algorithm and the decision of an oracle, who is the security guard of the forest look-out tower. ©2009 IEEE.