Browsing by Subject "Performance guarantees"
Now showing 1 - 3 of 3
- Results Per Page
- Sort Options
Item Open Access Adaptive ensemble learning with confidence bounds for personalized diagnosis(AAAI Press, 2016) Tekin, Cem; Yoon, J.; Van Der Schaar, M.With the advances in the field of medical informatics, automated clinical decision support systems are becoming the de facto standard in personalized diagnosis. In order to establish high accuracy and confidence in personalized diagnosis, massive amounts of distributed, heterogeneous, correlated and high-dimensional patient data from different sources such as wearable sensors, mobile applications, Electronic Health Record (EHR) databases etc. need to be processed. This requires learning both locally and globally due to privacy constraints and/or distributed nature of the multimodal medical data. In the last decade, a large number of meta-learning techniques have been proposed in which local learners make online predictions based on their locally-collected data instances, and feed these predictions to an ensemble learner, which fuses them and issues a global prediction. However, most of these works do not provide performance guarantees or, when they do, these guarantees are asymptotic. None of these existing works provide confidence estimates about the issued predictions or rate of learning guarantees for the ensemble learner. In this paper, we provide a systematic ensemble learning method called Hedged Bandits, which comes with both long run (asymptotic) and short run (rate of learning) performance guarantees. Moreover, we show that our proposed method outperforms all existing ensemble learning techniques, even in the presence of concept drift.Item Open Access Adaptive hierarchical space partitioning for online classification(IEEE, 2016) Kılıç, O. Fatih; Vanlı, N. D.; Özkan, H.; Delibalta, İ.; Kozat, Süleyman SerdarWe propose an online algorithm for supervised learning with strong performance guarantees under the empirical zero-one loss. The proposed method adaptively partitions the feature space in a hierarchical manner and generates a powerful finite combination of basic models. This provides algorithm to obtain a strong classification method which enables it to create a linear piecewise classifier model that can work well under highly non-linear complex data. The introduced algorithm also have scalable computational complexity that scales linearly with dimension of the feature space, depth of the partitioning and number of processed data. Through experiments we show that the introduced algorithm outperforms the state-of-the-art ensemble techniques over various well-known machine learning data sets.Item Open Access Average error in recovery of sparse signals and discrete fourier transform(IEEE, 2012-04) Özçelikkale, Ayça; Yüksel, S.; Özaktaş Haldun M.In compressive sensing framework it has been shown that a sparse signal can be successfully recovered from a few random measurements. The Discrete Fourier Transform (DFT) is one of the transforms that provide the best performance guarantees regardless of which components of the signal are nonzero. This result is based on the performance criterion of signal recovery with high probability. Whether the DFT is the optimum transform under average error criterion, instead of high probability criterion, has not been investigated. Here we consider this optimization problem. For this purpose, we model the signal as a random process, and propose a model where the covariance matrix of the signal is used as a measure of sparsity. We show that the DFT is, in general, not optimal despite numerous results that suggest otherwise. © 2012 IEEE.