Browsing by Subject "Object tracking"
Now showing 1 - 7 of 7
- Results Per Page
- Sort Options
Item Open Access Co-difference based object tracking algorithm for infrared videos(IEEE, 2016) Demir, H. Seçkin; Çetin, A. EnisThis paper presents a novel infrared (IR) object tracking algorithm based on the co-difference matrix. Extraction of co-difference features is similar to the well known covariance method except that the vector product operator is redefined in a multiplication-free manner. The new operator yields a computationally efficient implementation for real time object tracking applications. Experiments on an extensive set of IR image sequences indicate that the new method performs better than covariance tracking and other tracking algorithms without requiring any multiplication operations.Item Open Access Image sequence analysis for emerging interactive multimedia services-the European COST 211 framework(Institute of Electrical and Electronics Engineers, 1998-11) Alatan, A. A.; Onural, L.; Wollborn, M.; Mech, R.; Tuncel, E.; Sikora, T.Flexibility and efficiency of coding, content extraction, and content-based search are key research topics in the field of interactive multimedia. Ongoing ISO MPEG-4 and MPEG-7 activities are targeting standardization to facilitate such services. European COST Telecommunications activities provide a framework for research collaboration. COST 211 bis and COST 211 tcr activities have been instrumental in the definition and development of the ITU-T H.261 and H.263 standards for video-conferencing over ISDN and videophony over regular phone lines, respectively. The group has also contributed significantly to the ISO MPEG-4 activities. At present a significant effort of the COST 211 tcr group activities is dedicated toward image and video sequence analysis and segmentation - an important technological aspect for the success of emerging object-based MPEG-4 and MPEG-7 multimedia applications. The current work of COST 211 is centered around the test model, called the Analysis Model (AM). The essential feature of the AM is its ability to fuse information from different sources to achieve a high-quality object segmentation. The current information sources are the intermediate results from frame-based (still) color segmentation, motion vector based segmentation, and change-detection-based segmentation. Motion vectors, which form the basis for the motion vector based intermediate segmentation, are estimated from consecutive frames. A recursive shortest spanning tree (RSST) algorithm is used to obtain intermediate color and motion vector based segmentation results. A rule-based region processor fuses the intermediate results; a postprocessor further refines the final segmentation output. The results of the current AM are satisfactory; it is expected that there will be further improvements of the AM within the COST 211 project.Item Open Access Oscillatory synchronization model of attention to moving objects(Elsevier, 2012) Yilmaz, O.The world is a dynamic environment hence it is important for the visual system to be able to deploy attention on moving objects and attentively track them. Psychophysical experiments indicate that processes of both attentional enhancement and inhibition are spatially focused on the moving objects; however the mechanisms of these processes are unknown. The studies indicate that the attentional selection of target objects is sustained via a feedforward-feedback loop in the visual cortical hierarchy and only the target objects are represented in attention-related areas. We suggest that feedback from the attention-related areas to early visual areas modulates the activity of neurons; establishes synchronization with respect to a common oscillatory signal for target items via excitatory feedback, and also establishes de-synchronization for distractor items via inhibitory feedback. A two layer computational neural network model with integrate-and-fire neurons is proposed and simulated for simple attentive tracking tasks. Consistent with previous modeling studies, we show that via temporal tagging of neural activity, distractors can be attentively suppressed from propagating to higher levels. However, simulations also suggest attentional enhancement of activity for distractors in the first layer which represents neural substrate dedicated for low level feature processing. Inspired by this enhancement mechanism, we developed a feature based object tracking algorithm with surround processing. Surround processing improved tracking performance by 57% in PETS 2001 dataset, via eliminating target features that are likely to suffer from faulty correspondence assignments. © 2012 Elsevier Ltd.Item Open Access Real-time detection, tracking and classification of multiple moving objects in UAV videos(IEEE, 2017-11) Baykara, Hüseyin Can; Bıyık, Erdem; Gül, Gamze; Onural, Deniz; Öztürk, Ahmet Safa; Yıldız, İlkayUnnamed Aerial Vehicles (UAVs) are becoming increasingly popular and widely used for surveillance and reconnaissance. There are some recent studies regarding moving object detection, tracking, and classification from UAV videos. A unifying study, which also extends the application scope of such previous works and provides real-Time results, is absent from the literature. This paper aims to fill this gap by presenting a framework that can robustly detect, track and classify multiple moving objects in real-Time, using commercially available UAV systems and a common laptop computer. The framework can additionally deliver practical information about the detected objects, such as their coordinates and velocities. The performance of the proposed framework, which surpasses human capabilities for moving object detection, is reported and discussed.Item Open Access Semi-automatic video object segmentation(2000) Esen, ErsinContent-based iunetionalities form the core of the future multimedia applications. The new multimedia standard MPEG-4 provides a new form of interactivity with coded audio-visual data. The emerging standard MPEG-7 specifies a common description of various types of multimedia information to index the data for storage and retrieval. However, none of these standards specifies how to extract the content of the multimedia data. Video object segmentation addresses this task and tries to extract semantic objects from a scene. Two tyj)es of video object segmentation can be identified: unsupervised and supervised. In unsupervised méthods the user is not involved in any step of the process. In supervised methods the user is requested to supply additional information to increase the quality of the segmentation. The proposed weakly supervised still image segmentation asks the user to draw a scribble over what he defines as an object. These scribbles inititate the iterative method. .A.t each iteration the most similar regions are merged until the desired numljer of regions is reached. The proposed .segmentation method is inserted into the unsupervised COST211ter .A-ualysis Model (.A.M) for video object segmentation. The AM is modified to handh' the sujiervision. The new semi-automatic AM requires the user intei actimi for onl>· first frame of the video, then segmentation and object tracking is doin' automatically. The results indicate that the new semi-automatic AM constituK's a good tool for video oliject segmentation.Item Open Access Video object segmentation for interactive multimedia(1998) Ekmekçi, TolgaRecently, trends in video processing research have shifted from video compression to video analysis, due to the emerging standards MPEG-4 and MPEG-7. These standards will enable the users to interact with the objects in the audiovisual scene generated at the user’s end. However, neither of them prescribes how to obtain the objects. Many methods have been proposed for segmentation of video objects. One of the approaches is the “Analysis Model” (AM) of European COST-211 project. It is a modular approach to video object segmentation problem. Although AM performs acceptably in some cases, the results in many other cases are not good enough to be considered as semantic objects. In this thesis, a new tool is integrated and some modules are replaced by improved versions. One of the tools uses a block-based motion estimation technique to analyze the motion content within a scene, computes a motion activity parameter, and skips frames accordingly. Also introduced is a powerful motion estimation method which uses maximum a posteriori probability (MAP) criterion and Gibbs energies to obtain more reliable motion vectors and to calculate temporally unpredictable areas. To handle more complex motion in the scene, the 2-D affine motion model is added to the motion segmentation module, which employs only the translational model. The observed results indicate that the AM performance is improved substantially. The objects in the scene and their boundaries are detected more accurately, compared to the previous results.Item Open Access Visual object tracking in drone images with deep reinforcement learning(IEEE, 2021-05-05) Gözen, Derya; Özer, SedatThere is an increasing demand on utilizing camera equipped drones and their applications in many domains varying from agriculture to entertainment and from sports events to surveillance. In such drone applications, an essential and a common task is tracking an object of interest visually. Drone (or UAV) images have different properties when compared to the ground taken (natural) images and those differences introduce additional complexities to the existing object trackers to be directly applied on drone applications. Some important differences among those complexities include (i) smaller object sizes to be tracked and (ii) different orientations and viewing angles yielding different texture and features to be observed. Therefore, new algorithms trained on drone images are needed for the drone-based applications. In this paper, we introduce a deep reinforcement learning (RL) based single object tracker that tracks an object of interest in drone images by estimating a series of actions to find the location of the object in the next frame. This is the first work introducing a single object tracker using a deep RL-based technique for drone images. Our proposed solution introduces a novel reward function that aims to reduce the total number of actions taken to estimate the object's location in the next frame and also introduces a different backbone network to be used on low resolution images. Additionally, we introduce a set of new actions into the action library to better deal with the above-mentioned complexities. We compare our proposed solutions to a state of the art tracking algorithm from the recent literature and demonstrate up to 3.87 % improvement in precision and 3.6% improvement in IoU values on the VisDrone2019 data set. We also provide additional results on OTB-100 data set and show up to 3.15% improvement in precision on the OTB-100 data set when compared to the same previous state of the art algorithm. Lastly, we analyze the ability to handle some of the challenges faced during tracking, including but not limited to occlusion, deformation, and scale variation for our proposed solutions.