Browsing by Subject "Wavelet transform"
Now showing 1 - 16 of 16
- Results Per Page
- Sort Options
Item Open Access Compressive sensing based flame detection in infrared videos(IEEE, 2013) Günay, Osman; Çetin, A. EnisIn this paper, a Compressive Sensing based feature extraction algorithm is proposed for flame detection using infrared cameras. First, bright and moving regions in videos are detected. Then the videos are divided into spatio-temporal blocks and spatial and temporal feature vectors are exctracted from these blocks. Compressive Sensing is used to exctract spatial feature vectors. Compressed measurements are obtained by multiplying the pixels in the block with the sensing matrix. A new method is also developed to generate the sensing matrix. A random vector generated according to standard Gaussian distribution is passed through a wavelet transform and the resulting matrix is used as the sensing matrix. Temporal features are obtained from the vector that is formed from the difference of mean intensity values of the frames in two neighboring blocks. Spatial feature vectors are classified using Adaboost. Temporal feature vectors are classified using hidden Markov models. To reduce the computational cost only moving and bright regions are classified and classification is performed at specified intervals instead of every frame. © 2013 IEEE.Item Open Access Computer vision based method for real-time fire and flame detection(Elsevier BV, 2006-01-01) Töreyin, B. U.; Dedeoǧlu, Y.; Güdükbay, Uğur; Çetin, A. EnisThis paper proposes a novel method to detect fire and/or flames in real-time by processing the video data generated by an ordinary camera monitoring a scene. In addition to ordinary motion and color clues, flame and fire flicker is detected by analyzing the video in the wavelet domain. Quasi-periodic behavior in flame boundaries is detected by performing temporal wavelet transform. Color variations in flame regions are detected by computing the spatial wavelet transform of moving fire-colored regions. Another clue used in the fire detection algorithm is the irregularity of the boundary of the fire-colored region. All of the above clues are combined to reach a final decision. Experimental results show that the proposed method is very successful in detecting fire and/or flames. In addition, it drastically reduces the false alarms issued to ordinary fire-colored moving objects as compared to the methods using only motion and color clues. © 2005 Elsevier B.V. All rights reserved.Item Open Access Contact-free measurement of respiratory rate using infrared and vibration sensors(Elsevier BV, 2015) Erden, F.; Alkar, A. Z.; Çetin, A. EnisRespiratory rate is an essential parameter in many practical applications such as apnea detection, patient monitoring, and elderly people monitoring. In this paper, we describe a novel method and a contact-free multi-modal system which is capable of detecting human breathing activity. The multimodal system, which uses both differential pyro-electric infrared (PIR) and vibration sensors, can also estimate the respiratory rate. Vibration sensors pick up small vibrations due to the breathing activity. Similarly, PIR sensors pick up the thoracic movements. Sensor signals are sampled using a microprocessor board and analyzed on a laptop computer. Sensor signals are processed using wavelet analysis and empirical mode decomposition (EMD). Since breathing is almost periodic, a new multi-modal average magnitude difference function (AMDF) is used to detect the periodicity and the period in the processed signals. By fusing the data of two different types of sensors we achieve a more robust and reliable contact-free human breathing activity detection system compared to systems using only one specific type of sensors.Item Open Access Edge projections for eye localization(SPIE - International Society for Optical Engineering, 2008) Turkan, M.; Pardas, M.; Çetin, A. EnisAn algorithm for human-eye localization in images is presented for faces with frontal pose and upright orientation. A given face region is filtered by a highpass wavelet-transform filter. In this way, edges of the region are highlighted, and a caricature-like representation is obtained. Candidate points for each eye are detected after analyzing horizontal projections and profiles of edge regions in the highpass-filtered image. All the candidate points are then classified using a support vector machine. Locations of each eye are estimated according to the most probable ones among the candidate points. It is experimentally observed that our eye localization method provides promising results for image-processing applications.Item Open Access Hand gesture based remote control system using infrared sensors and a camera(Institute of Electrical and Electronics Engineers, 2014) Erden, F.; Çetin, A.In this paper, a multimodal hand gesture detection and recognition system using differential Pyroelectric Infrared (PIR) sensors and a regular camera is described. Any movement within the viewing range of the differential PIR sensors are first detected by the sensors and then checked if it is due to a hand gesture or not by video analysis. If the movement is due to a hand, one-dimensional continuous-time signals extracted from the PIR sensors are used to classify/recognize the hand movements in real-time. Classification of different hand gestures by using the differential PIR sensors is carried out by a new winner-takeall (WTA) hash based recognition method. Jaccard distance is used to compare the WTA hash codes extracted from 1-D differential infrared sensor signals. It is experimentally shown that the multimodal system achieves higher recognition rates than the system based on only the on/off decisions of the analog circuitry of the PIR sensors.Item Open Access Heart sound segmentation using signal processing methods(2015) Şahin, DevrimHeart murmurs are pathological heart sounds that originate from blood flowing with abnormal turbulence due to physiological defects of the heart, and are the prime indicator of many heart-related diseases. Murmurs can be diagnosed via auscultation; that is, by listening with a stethoscope. However, manual detection and classification of murmur requires clinical expertise and is highly prone to misclassification. Although automated classification algorithms exist for this purpose; they heavily depend on feature extraction from ‘segmented’ heart sound waveforms. Segmentation in this context refers to detecting and splitting cardiac cycles. However, heart sound signal is not a stationary signal; and typically has a low signal-to-noise ratio, which makes it very difficult to segment using no external information but the signal itself. Most of the commercial systems require an external electrocardiography (ECG) signal to determine S1 and S2 peaks, but ECG is not as widely available as stethoscopes. Although algorithms that provide segmentation using sound alone exist, a proper comparison between these algorithms on a common dataset is missing. We propose several modifications to many of these algorithms, as well as an evaluation method that allows a unified comparison of all these approaches. We have tested each combination of algorithms on a real data set [1], which also provides manual annotations as ground truth. We also propose an ensemble of several methods, and a heuristic for which algorithm’s output to use. Whereas tested algorithms report up to 62% accuracy, our ensemble method reports a 75% success rate. Finally, we created a tool named UpBeat to enable manual segmentation of heart sounds, and construction of a ground truth dataset. UpBeat is a starting medium for auscultation segmentation, time-domain based feature extraction and evaluation; which has automatic segmentation capabilities, as well as a minimalistic drag-and-drop interface which allows manual annotation of S1 and S2 peaks.Item Open Access Human eye localization using edge projections(Institute for Systems and Technologies of Information, Control and Communication, 2007) Türkan, Mehmet; Pardas, M.; Çetin, A. EnisIn this paper, a human eye localization algorithm in images and video is presented for faces with frontal pose and upright orientation. A given face region is filtered by a high-pass filter of a wavelet transform. In this way, edges of the region are highlighted, and a caricature-like representation is obtained. After analyzing horizontal projections and profiles of edge regions in the high-pass filtered image, the candidate points for each eye are detected. All the candidate points are then classified using a support vector machine based classifier. Locations of each eye are estimated according to the most probable ones among the candidate points. It is experimentally observed that our eye localization method provides promising results for both image and video processing applications.Item Open Access Image processing algorithms for histopathological images(2016-03) Oğuz, OğuzhanConventionally, a pathologist examines cancer cell morphologies under microscope. This process takes a lot of time and is subject to human mistakes. Computer aided diagnosis (CAD) systems and modules aim to help pathologists in their work to decrease the time consumption and the human mistakes. This thesis proposes a CAD module and algorithms which assist the pathologist in segmentation, detection and the classi cation problems in histopatholgic images. A multi-resolution super-pixel based segmentation algorithm is developed to measure the cell size, count the number of cells and track the motion of cells in Mesenchymal Stem Cell (MSC) images. The proposed algorithm is compared with Simple Linear Iterative Clustering (SLIC) algorithm. It is experimentally observed that in the segmentation stage, the cell detection rate is increased by 7% and the false alarm is decreased by 5%. In addition to this, two novel decision rules for merging similar neighboring super-pixels are proposed. One dimensional version of the Scale Invariant Feature Transform (SIFT) based merging algorithm is developed and applied to the histograms of the neighboring super-pixels to determine the similar regions. It is also shown that the merging process can be made with the use of wavelets. Moreover, it is shown that region covariance and codi erence matrices can be used in detection of cancer stem cells (CSC) and a CAD module for the CSC detection in liver cancer tissue images are developed. The system locates CSCs in CD13 stained liver tissue images. The method has an online learning approach which improves the accuracy of detection. It is experimentally shown that, applying the proposed approach with the user guidance, increases the overall detection quality and accuracy up to 25% compared to using region descriptors alone. Also, the proposed module is compared with the similar plug-ins of ImageJ and Fiji. It is shown that, when the similar features are used, the implemented module achieves approximately 20% better classi cation results compared to the plug-ins of Imagej and Fiji. Furthermore, the proposed 1-D SIFT algorithm is expanded and used in classi cation of the cancer tissues images stained with Hematoxylin and Eosin (H&E) stain, which is a cost e ective routine compared to the immunohistochemistry (IHC) procedure. The 1-D SIFT algorithm is able to classify healthy and cancerous tissue images with up to 91% accuracy in H&E stained images in our data set.Item Open Access Moving object detection using adaptive subband decomposition and fractional lower-order statistics in video sequences(Elsevier, 2002) Bagci, A. M.; Yardimci, Y.; Çetin, A. EnisIn this paper, a moving object detection method in video sequences is described. In the first step, the camera motion is eliminated using motion compensation. An adaptive subband decomposition structure is then used to analyze the motion compensated image. In the "low-high" and "high-low" subimages moving objects appear as outliers and they are detected using a statistical detection test based on fractional lower-order statistics. It turns out that the distribution of the subimage pixels is almost Gaussian in general. On the other hand, at the object boundaries the distribution of the pixels in the subimages deviates from Gaussianity due to the existence of outliers. By detecting the regions containing outliers the boundaries of the moving objects are estimated. Simulation examples are presented. © 2002 Elsevier Science B.V. All rights reserved.Item Open Access New methods for robust speech recognition(1995) Erzin, EnginNew methods of feature extraction, end-point detection and speech enhcincement are developed for a robust speech recognition system. The methods of feature extraction and end-point detection are based on wavelet analysis or subband analysis of the speech signal. Two new sets of speech feature parameters, SUBLSF’s and SUBCEP’s, are introduced. Both parameter sets are based on subband analysis. The SUBLSF feature parameters are obtained via linear predictive analysis on subbands. These speech feature parameters can produce better results than the full-band parameters when the noise is colored. The SUBCEP parameters are based on wavelet analysis or equivalently the multirate subband analysis of the speech signal. The SUBCEP parameters also provide robust recognition performance by appropriately deemphasizing the frequency bands corrupted by noise. It is experimentally observed that the subband analysis based feature parameters are more robust than the commonly used full-band analysis based parameters in the presence of car noise. The a-stable random processes can be used to model the impulsive nature of the public network telecommunication noise. Adaptive filtering are developed for Q-stable random processes. Adaptive noise cancelation techniques are used to reduce the mismacth between training and testing conditions of the recognition system over telephone lines. Another important problem in isolated speech recognition is to determine the boundaries of the speech utterances or words. Precise boundary detection of utterances improves the performance of speech recognition systems. A new distance measure based on the subband energy levels is introduced for endpoint detection.Item Open Access Nonrectangular wavelets for multiresolution mesh analysis and compression(SPIE, 2006) Köse, Kıvanç; Çetin, A. Enis; Güdükbay, Uğur; Onural, LeventWe propose a new Set Partitioning In Hierarchical Trees (SPIHT) based mesh compression framework. The 3D mesh is first transformed to 2D images on a regular grid structure. Then, this image-like representation is wavelet transformed and SPIHT is applied on the wavelet domain data. The method is progressive because the resolution of the reconstructed mesh can be changed by varying the length of the ID data stream created by SPIHT algorithm. Nearly perfect reconstruction is possible if full length of 1D data is received.Item Open Access Small moving object detection using adaptive subband decomposition and fractional lower order statistics in video sequences(2001-07-08) Bağci, A.Murat; Yardımcı, Y.; Çetin, A. EnisIn this paper, a small moving object method detection method in video sequences is described. In the first step, the camera motion is eliminated using motion compensation. An adaptive subband decomposition structure is then used to analyze the motion compensated image. In the highband subimages moving objects appear as outliers and they are detected using a statistical detection test based on lower order statistics. It turns out that in general, the distribution of the residual error image pixels is almost Gaussian. On the other hand, the distribution of the pixels in the residual image deviates from Gaussianity in the existence of outliers. By detecting the regions containing outliers the boundaries of the moving objects are estimated. Simulation examples are presented.Item Open Access Video processing algorithms for wildfire surveillance(2015-05) Günay, OsmanWe propose various image and video processing algorithms for wild re surveillance. The proposed methods include; classi er fusion, online learning, real-time feature extraction, image registration and optimization. We develop an entropy functional based online classi er fusion framework. We use Bregman divergences as the distance measure of the projection operator onto the hyperplanes describing the output decisions of classi ers. We test the performance of the proposed system in a wild re detection application with stationary cameras that scan prede ned preset positions. In the second part of this thesis, we investigate di erent formulations and mixture applications for passive-aggressive online learning algorithms. We propose a classi er fusion method that can be used to increase the performance of multiple online learners or the same learners trained with di erent update parameters. We also introduce an aerial wild re detection system to test the real-time performance of the analyzed algorithms. In the third part of the thesis we propose a real-time dynamic texture recognition method using random hyperplanes and deep neural networks. We divide dynamic texture videos into spatio-temporal blocks and extract features using local binary patterns (LBP). We reduce the computational cost of the exhaustive LBP method by using randomly sampled subset of pixels in the block. We use random hyperplanes and deep neural networks to reduce the dimensionality of the nal feature vectors. We test the performance of the proposed method in a dynamic texture database. We also propose an application of the proposed method in real-time detection of ames in infrared videos. Using the same features we also propose a fast wild re detection system using pan-tilt-zoom cameras and panoramic background subtraction. We use a hybrid method consisting of speeded-up robust features and mutual information to register consecutive images and form the panorama. The next step for multi-modal surveillance applications is the registration of images obtained with di erent devices. We propose a multi-modal image registration algorithm for infrared and visible range cameras. A new similarity measure is described using log-polar transform and mutual information to recover rotation and scale parameters. Another similarity measure is introduced using mutual information and redundant wavelet transform to estimate translation parameters. The new cost function for translation parameters is minimized using a novel lifted projections onto convex sets method.Item Open Access VOC gas leak detection using pyro-electric infrared sensors(IEEE, 2010) Erden, Fatih; Soyer, E. B.; Toreyin, B. U.; Çetin, A. EnisIn this paper, we propose a novel method for detecting and monitoring Volatile Organic Compounds (VOC) gas leaks by using a Pyro-electric (or Passive) Infrared (PIR) sensor whose spectral range intersects with the absorption bands of VOC gases. A continuous time analog signal is obtained from the PIR sensor. This signal is discretized and analyzed in real time. Feature parameters are extracted in wavelet domain and classified using a Markov Model (MM) based classifier. Experimental results are presented. ©2010 IEEE.Item Open Access Volatile organic compound plume detection using wavelet analysis of video(IEEE, 2008-10) Töreyin, B. Uğur; Çetin, A. EnisA video based method to detect volatile organic compounds (VOC) leaking out of process equipments used in petrochemical refineries is developed. Leaking VOC plume from a damaged component causes edges present in image frames loose their sharpness. This leads to a decrease in the high frequency content of the image. The background of the scene is estimated and decrease of high frequency energy of the scene is monitored using the spatial wavelet transforms of the current and the background images. Plume regions in image frames are analyzed in low-band sub-images, as well. Image frames are compared with their corresponding low-band images. A maximum likelihood estimator (MLE) for adaptive threshold estimation is also developed in this paper. © 2008 IEEE.Item Open Access Wavelet based flickering flame detector using differential PIR sensors(Elsevier, 2012-07-06) Erden, F.; Toreyin, B. U.; Soyer, E. B.; Inac, I.; Gunay, O.; Kose, K.; Çetin, A. EnisA Pyro-electric Infrared (PIR) sensor based flame detection system is proposed using a Markovian decision algorithm. A differential PIR sensor is only sensitive to sudden temperature variations within its viewing range and it produces a time-varying signal. The wavelet transform of the PIR sensor signal is used for feature extraction from sensor signal and wavelet parameters are fed to a set of Markov models corresponding to the flame flicker process of an uncontrolled fire, ordinary activity of human beings and other objects. The final decision is reached based on the model yielding the highest probability among others. Comparative results show that the system can be used for fire detection in large rooms.