Browsing by Subject "Information management"
Now showing 1 - 20 of 23
- Results Per Page
- Sort Options
Item Open Access ALACA: a platform for dynamic alarm collection and alert notification in network management systems(John Wiley and Sons Ltd., 2017) Solmaz, S. E.; Gedik, B.; Ferhatosmanoğlu, H.; Sözüer, S.; Zeydan, E.; Etemoğlu, Ç. Ö.Mobile network operators run Operations Support Systems that produce vast amounts of alarm events. These events can have different significance levels and domains and also can trigger other ones. Network operators face the challenge to identify the significance and root causes of these system problems in real time and to keep the number of remedial actions at an optimal level, so that customer satisfaction rates can be guaranteed at a reasonable cost. In this paper, we propose a scalable streaming alarm management system, referred to as Alarm Collector and Analyzer, that includes complex event processing and root cause analysis. We describe a rule mining and root cause analysis solution for alarm event correlation and analyses. The solution includes a dynamic index for matching active alarms, an algorithm for generating candidate alarm rules, a sliding window–based approach to save system resources, and a graph-based solution to identify root causes. Alarm Collector and Analyzer is used in the network operation center of a major mobile telecom provider. It helps operators to enhance the design of their alarm management systems by allowing continuous analysis of data and event streams and predict network behavior with respect to potential failures by using the results of root cause analysis. We present experimental results that provide insights on performance of real-time alarm data analytics systems. Copyright © 2017 John Wiley & Sons, Ltd.Item Open Access An archiving model for a hierarchical information storage environment(Elsevier, 2000) Moinzadeh, K.; Berk, E.We consider an archiving model for a database consisting of secondary and tertiary storage devices in which the query rate for a record declines as it ages. We propose a `dynamic' archiving policy based on the number of records and the age of the records in the secondary device. We analyze the cases when the number of new records inserted in the system over time are either constant or follow a Poisson process. For both scenarios, we characterize the properties of the policy parameters and provide optimization results when the objective is to minimize the average record retrieval times. Furthermore, we propose a simple heuristic method for obtaining near-optimal policies in large databases when the record query rate declines exponentially with time. The e ectiveness of the heuristic is tested via a numerical experiment. Finally, we examine the behavior of performance measures such as the average record retrieval time and the hit rate as system parameters are varied.Item Open Access Authorship attribution: performance of various features and classification methods(IEEE, 2007-11) Bozkurt, İlker Nadi; Bağlıoğlu, Özgür; Uyar, ErkanAuthorship attribution is the process of determining the writer of a document. In literature, there are lots of classification techniques conducted in this process. In this paper we explore information retrieval methods such as tf-idf structure with support vector machines, parametric and nonparametric methods with supervised and unsupervised (clustering) classification techniques in authorship attribution. We performed various experiments with articles gathered from Turkish newspaper Milliyet. We performed experiments on different features extracted from these texts with different classifiers, and combined these results to improve our success rates. We identified which classifiers give satisfactory results on which feature sets. According to experiments, the success rates dramatically changes with different combinations, however the best among them are support vector classifier with bag of words, and Gaussian with function words. ©2007 IEEE.Item Open Access Deploy-DDS: Tool framework for supporting deployment architecture of data distribution service based systems(ACM, 2014-08) Çelik, T.; Köksal, O.; Tekinerdoğan, BedirData Distribution Service (DDS) is the Object Management Group's (OMG) new standard middleware after Common Object Request Broker Architecture (CORBA), which is becoming increasingly popular. One of the important problems in DDS Based Software Systems is the deployment configuration of DDS modules to the physical resources. In general, this can be done in many different ways whereby each deployment alternative will perform differently. Currently, the deployment configuration is decided after the coding phase and usually performed manually. For large configurations, finding the feasible deployment might require serious rework with costly and time consuming iterations. In this paper, we present the tool Deploy-DDS to support the selection and generation of deployment architectures of DDS based systems. The tool can be used to perform an evaluation during the design phase and generate the selected feasible configuration. © 2014 Authors.Item Open Access Free riding in peer-to-peer networks(Institute of Electrical and Electronics Engineers, 2009) Karakaya, M.; Korpeoglu, I.; Ulusoy, ÖzgürFree riding in peer-to-peer (P2P) networks poses a serious threat to their proper operation. Here, the authors present a variety of approaches developed to overcome this problem. They introduce several unique aspects of P2P networks and discuss free riding's effects on P2P services. They categorize proposed solutions and describe each category's important features and implementation issues together with some sample solutions. They also discuss open issues, including common attacks and security considerations. © 2009 IEEE.Item Open Access Generalized ID-based ElGamal signatures(IEEE, 2007-11) Kalkan, Said; Kaya, Kamer; Selçuk, Ali AydınID-based cryptography has been a very active area of research in cryptography since bilinear pairings were introduced as a cryptographic tool, and there have been many proposals for ID-based signatures recently. In this paper, we introduce the concept of generalized ID-based ElGamal signatures and show that most of the proposed ID-based signature schemes in the literature are special instances of this generalized scheme. We also obtain numerous new signatures from this generalized scheme which have not been proposed before. ©2007 IEEE.Item Open Access IBM streams processing language: analyzing big data in motion(I B M Corp., 2013-05-17) Hirzel M.; Andrade, H.; Gedik, B.; Jacques-Silva, R.; Khandekar, R.; Kumar, V.; Mendell, M.; Nasgaard, H.; Schneider S.; Soule´, R.; Wu, K. L.The IBM Streams Processing Language (SPL) is the programming language for IBM InfoSphere® Streams, a platform for analyzing Big Data in motion. By “Big Data in motion,” we mean continuous data streams at high data-transfer rates. InfoSphere Streams processes such data with both high throughput and short response times. To meet these performance demands, it deploys each application on a cluster of commodity servers. SPL abstracts away the complexity of the distributed system, instead exposing a simple graph-of-operators view to the user. SPL has several innovations relative to prior streaming languages. For performance and code reuse, SPL provides a code-generation interface to C++ and Java®. To facilitate writing well-structured and concise applications, SPL provides higher-order composite operators that modularize stream sub-graphs. Finally, to enable static checking while exposing optimization opportunities, SPL provides a strong type system and user-defined operator models. This paper provides a language overview, describes the implementation including optimizations such as fusion, and explains the rationale behind the language design.Item Open Access Implications of non-volatile memory as primary storage for database management systems(IEEE, 2017) Mustafa, Naveed Ul; Armejach, A.; Öztürk, Özcan; Cristal, A.; Unsal, O. S.Traditional Database Management System (DBMS) software relies on hard disks for storing relational data. Hard disks are cheap, persistent, and offer huge storage capacities. However, data retrieval latency for hard disks is extremely high. To hide this latency, DRAM is used as an intermediate storage. DRAM is significantly faster than disk, but deployed in smaller capacities due to cost and power constraints, and without the necessary persistency feature that disks have. Non-Volatile Memory (NVM) is an emerging storage class technology which promises the best of both worlds. It can offer large storage capacities, due to better scaling and cost metrics than DRAM, and is non-volatile (persistent) like hard disks. At the same time, its data retrieval time is much lower than that of hard disks and it is also byte-addressable like DRAM. In this paper, we explore the implications of employing NVM as primary storage for DBMS. In other words, we investigate the modifications necessary to be applied on a traditional relational DBMS to take advantage of NVM features. As a case study, we have modified the storage engine (SE) of PostgreSQL enabling efficient use of NVM hardware. We detail the necessary changes and challenges such modifications entail and evaluate them using a comprehensive emulation platform. Results indicate that our modified SE reduces query execution time by up to 40% and 14.4% when compared to disk and NVM storage, with average reductions of 20.5% and 4.5%, respectively. © 2016 IEEE.Item Open Access Induction of logical relations based on specific generalization of strings(IEEE, 2007-11) Uzun, Yasin; Çiçekli, İlyasLearning logical relations from examples expressed as first order facts has been studied extensively by the Inductive Logic Programming research. Learning with positive-only data may cause over generalization of examples leading to inconsistent resulting hypotheses. A learning heuristic inferring specific generalization of strings based on unique match sequences is shown to be capable of learning predicates with string arguments. This paper describes an inductive learner based on the idea of specific generalization of strings, and the given clauses are generalized by considering the background knowledge. ©2007 IEEE.Item Open Access Item Open Access Multi-item quick response system with budget constraint(2012) Serel, D. A.Quick response mechanisms based on effective use of up-to-date demand information help retailers to reduce their inventory management costs. We formulate a single-period inventory model for multiple products with dependent (multivariate normal) demand distributions and a given overall procurement budget. After placing orders based on an initial demand forecast, new market information is gathered and demand forecast is updated. Using this more accurate second forecast, the retailer decides the total stocking level for the selling season. The second order is based on an improved demand forecast, but it also involves a higher unit supply cost. To determine the optimal ordering policy, we use a computational procedure that entails solving capacitated multi-item newsboy problems embedded within a dynamic programming model. Various numerical examples illustrate the effects of demand variability and financial constraint on the optimal policy. It is found that existence of a budget constraint may lead to an increase in the initial order size. It is also observed that as the budget available decreases, the products with more predictable demand make up a larger share of the procurement expenditure.Item Open Access Optimal timing of project control points(Elsevier, 2000) Raz, T.; Erel, E.The project control cycle consists of measuring the status of the project, comparing to the plan, analysis of the deviations, and implementing any appropriate corrective actions. We present an analytical framework for determining the optimal timing of project control points throughout the life cycle of the project. Our approach is based on maximizing the amount of information generated by the control points, which depends on the intensity of the activities carried out since the last control point and on the time elapsed since their execution. The optimization problem is solved with a dynamic programming approach. We report the results of numerical experimentation with the model involving different types of activity intensity profiles and several levels of information loss. For each combination, we compared the optimal amount of information to the amount of information obtained with two simpler policies: control at equal time intervals, and control at equal activity contents intervals. We also investigated the effect of adding more control points on the amount of information generated.Item Open Access Parallel preconditioners for solutions of dense linear systems with tens of millions of unknowns(2007-11) Malas, Tahir; Ergül, Özgür; Gürel, LeventWe propose novel parallel preconditioning schemes for the iterative solution of integral equation methods. In particular, we try to improve convergence rate of the ill-conditioned linear systems formulated by the electric-field integral equation, which is the only integral-equation formulation for targets having open surfaces. For moderate-size problems, iterative solution of the near-field system enables much faster convergence compared to the widely used sparse approximate inverse preconditioner. For larger systems, we propose an approximation strategy to the multilevel fast multipole algorithm (MLFMA) to be used as a preconditioner. Our numerical experiments reveal that this scheme significantly outperforms other preconditioners. With the combined effort of effective preconditioners and an efficiently parallelized MLFMA, we are able to solve targets with tens of millions of unknowns, which are the largest problems ever reported in computational electromagnetics. ©2007 IEEE.Item Open Access PetaShare: A reliable, efficient and transparent distributed storage management system(2011) Kosar, T.; Akturk I.; Balman, M.; Wang X.Modern collaborative science has placed increasing burden on data management infrastructure to handle the increasingly large data archives generated. Beside functionality, reliability and availability are also key factors in delivering a data management system that can efficiently and effectively meet the challenges posed and compounded by the unbounded increase in the size of data generated by scientific applications. We have developed a reliable and efficient distributed data storage system, PetaShare, which spans multiple institutions across the state of Louisiana. At the back-end, PetaShare provides a unified name space and efficient data movement across geographically distributed storage sites. At the front-end, it provides light-weight clients the enable easy, transparent and scalable access. In PetaShare, we have designed and implemented an asynchronously replicated multi-master metadata system for enhanced reliability and availability, and an advanced buffering system for improved data transfer performance. In this paper, we present the details of our design and implementation, show performance results, and describe our experience in developing a reliable and efficient distributed data management system for data-intensive science. © 2011 - IOS Press and the authors. All rights reserved.Item Open Access PocketDrive: A system for mobile control of desktop PC and its applications using PDAs(IEEE, 2007-11) Yıldırım, Yenel; Körpeoğlu, İbrahimToday, consumer electronic devices and PCs are inevitable parts of our daily life. Controlling those devices remotely is an important aspect of the technology. We have already universal remote control devices for controlling consumer electronic devices. Similarly, we may control our desktop and laptop PCs and their applications remotely via portable and smaller computers like PDAs and Pocket PCs. This paper presents a system and its architecture that enable a wireless-enabled PDA to control a PC and its applications remotely over a 802.11 or Bluetooth link. With such a system running on a PDA, a user can start, run and control PC applications from any location that is reachable via 802.11 link. This enables flexibility, ease of use, and freedom for the user of PC applications. ©2007 IEEE.Item Open Access Quality assessment in the blog space(2010) Schaal, M.; Fidan G.; Müller, R.M.; Dagli O.Purpose: The purpose of this paper is the presentation of a new method for blog quality assessment. The method uses the temporal sequence of link creation events between blogs as an implicit source for the collective tacit knowledge of blog authors about blog quality. Design/methodology/approach: The blog data are processed by the novel method for the assessment of blog quality. The results are compared to Google Page Rank with respect to the Gold Standard, the BlogRazzi Bookmark Rank. Findings: The method is similar or better than Google Page Rank with respect to the chosen Gold Standard. Originality/value: The major contribution of this paper is the introduction of a novel method for blog quality assessment. Even though its superiority to other and more established methods cannot be proven in the context of this limited study, it enriches the toolset available for blog quality assessment and may become important for a deeper understanding of organizational learning. © Emerald Group Publishing Limited.Item Open Access Research issues in peer-to-peer data management(IEEE, 2007-11) Ulusoy, ÖzgürData management in Peer-to-Peer (P2P) systems is a complicated and challenging issue due to the scale of the network and highly transient population of peers. In this paper, we identify important research problems in P2P data management, and describe briefly some methods that have appeared in the literature addressing those problems. We also discuss some open research issues and directions regarding data management in P2P systems. ©2007 IEEE.Item Open Access Segmenting and labeling query sequences in a multidatabase environment(Springer, Berlin, Heidelberg, 2011) Acar, Aybar C.; Motro, A.When gathering information from multiple independent data sources, users will generally pose a sequence of queries to each source, combine (union) or cross-reference (join) the results in order to obtain the information they need. Furthermore, when gathering information, there is a fair bit of trial and error involved, where queries are recursively refined according to the results of a previous query in the sequence. From the point of view of an outside observer, the aim of such a sequence of queries may not be immediately obvious. We investigate the problem of isolating and characterizing subsequences representing coherent information retrieval goals out of a sequence of queries sent by a user to different data sources over a period of time. The problem has two sub-problems: segmenting the sequence into subsequences, each representing a discrete goal; and labeling each query in these subsequences according to how they contribute to the goal. We propose a method in which a discriminative probabilistic model (a Conditional Random Field) is trained with pre-labeled sequences. We have tested the accuracy with which such a model can infer labels and segmentation on novel sequences. Results show that the approach is very accurate (> 95% accuracy) when there are no spurious queries in the sequence and moderately accurate even in the presence of substantial noise (∼70% accuracy when 15% of queries in the sequence are spurious). © 2011 Springer-Verlag.Item Open Access Solution of large-scale scattering problems with the multilevel fast multipole algorithm parallelized on distributed-memory architectures(IEEE, 2007) Ergül, Özgür; Gürel, LeventWe present the solution of large-scale scattering problems involving three-dimensional closed conducting objects with arbitrary shapes. With an efficient parallelization of the multilevel fast multipole algorithm on relatively inexpensive computational platforms using distributed-memory architectures, we perform the iterative solution of integral-equation formulations that are discretized with tens of millions of unknowns. In addition to canonical problems, we also present the solution of real-life problems involving complicated targets with large dimensions.Item Open Access Threshold broadcast encryption with reduced complexity(IEEE, 2007-11) Kaşkaloǧlu, K.; Kaya, Kamer; Selçuk, Ali AydınThreshold Broadcast Encryption (TBE) is a promising extension of threshold cryptography with its advantages over traditional threshold cryptosystems, such as eliminating the need of a trusted party, the ability of setting up the system by individual users independently and the ability of choosing the threshold parameter and the group of privileged receivers at the time of encryption. An ElGamal-based solution for TBE was proposed by Ghodosi et al. In this paper, we propose an improved ElGamal-based TBE scheme with reduced transmission cost. ©2007 IEEE.