• HOME
  • IEEE Projects
    • IEEE Projects 2017 Dot Net Projects
    • IEEE Projects 2017 Java Projects
    • IEEE Projects 2016 Dot Net Projects
    • IEEE Projects 2016 Java Projects
    • IEEE Projects 2015
    • IEEE Projects 2014
      • IEEE 2014 Java Projects
        • IEEE Projects 2014 For Cse in Data Mining Java
        • IEEE Projects 2014 For Cse in cloud computing Java
        • IEEE Projects 2014 For Cse in Image Processing Java
        • IEEE Projects 2014 For Cse in Mobile Computing Java
        • IEEE Projects 2014 For Cse in Networking Java
        • IEEE Projects 2014 For Cse in Network Security Java
        • IEEE Projects 2014 For Cse in Software Engineering Java
      • IEEE 2014 Dotnet Projects
        • IEEE Projects 2014 For Cse in Data Mining Dotnet
        • IEEE Projects 2014 For Cse in Cloud Computing Dotnet
        • IEEE Projects 2014 For Cse in Netwoking Dotnet
        • IEEE Projects 2014 For Cse in Netwok Security Dotnet
    • IEEE Projects 2013
      • IEEE 2013 JAVA Projects
      • IEEE 2013 Dotnet Projects
    • IEEE Projects 2012
      • IEEE 2012 JAVA Projects
      • IEEE 2012 Dotnet Projects
    • IEEE Projects 2011
      • IEEE 2011 JAVA Projects
      • IEEE 2011 Dotnet Projects
    • IEEE Projects 2010
  • Power Electronics Projects
    • IEEE Projects 2015 For Power Electronics
    • IEEE Projects 2014 For Power Electronics
    • IEEE 2013 Power Electronics Projects
  • EMBEDDED Projects
    • IEEE Projects 2015 For Embedded Systems
    • IEEE 2013 Embedded Projects
  • Matlab Projects
    • IEEE 2013 Image Processing Projects
    • IEEE 2013 Power Electronics Projects
    • IEEE 2013 Communication Projects
  • NS2 Projects

Phd Projects | IEEE Project | IEEE Projects 2020-19 in Trichy & Chennai

IEEE Projects Trichy, Best IEEE Project Centre Chennai, Final Year Projects in Trichy - We Provide IEEE projects 2018 - 2019 , IEEE 2018 Java Projects for M.E/M.Tech, IEEE 2018 Dot net Projects for B.E/B.Tech, IEEE 2018 Power electronics Projects Engineering & Diploma Students, Matlab, Embedded, NS2 Projects
  • HOME
  • IEEE 2017 DOT NET PROJECT TITLES
  • IEEE 2017 JAVA PROJECT TITLES
  • CONTACT US

IEEE Projects 2014 For Cse in Data Mining Dotnet

TECHNOLOGY: DOTNET

DOMAIN: DATA MINING

 S. No.  IEEE TITLE                      ABSTRACT IEEE YEAR
1. A Group Incremental Approach to Feature Selection Applying Rough Set Technique Many real data increase dynamically in size. This phenomenon occurs in several fields including economics, population studies, and medical research. As an effective and efficient mechanism to deal with such data, incremental technique has been proposed in the literature and attracted much attention, which stimulates the result in this paper. When a group of objects are added to a decision table, we first introduce incremental mechanisms for three representative information entropies and then develop a group incremental rough feature selection algorithm based on information entropy. When multiple objects are added to a decision table, the algorithm aims to find the new feature subset in a much shorter time. Experiments have been carried out on eight UCI data sets and the experimental results show that the algorithm is effective and efficient.        2014
2. Consensus-Based Ranking of Multivalued Objects: A Generalized Borda Count Approach Abstract—In this paper, we tackle a novel problem of ranking multivalued objects, where an object has multiple instances in a multidimensional space, and the number of instances per object is not fixed. Given an ad hoc scoring function that assigns a score to a multidimensional instance, we want to rank a set of multivalued objects. Different from the existing models of ranking uncertain and probabilistic data, which model an object as a random variable and the instances of an object are assumed exclusive, we have to capture the coexistence of instances here. To tackle the problem, we advocate the semantics of favoring widely preferred objects instead of majority votes, which is widely used in many elections and competitions. Technically, we borrow the idea from Borda Count (BC), a well-recognized method in consensus-based voting systems. However, Borda Count cannot handle multivalued objects of inconsistent cardinality, and is costly to evaluate top k queries on large multidimensional data sets. To address the challenges, we extend and generalize Borda Count to quantile-based Borda Count, and develop efficient computational methods with comprehensive cost analysis. We present case studies on real data sets to demonstrate the effectiveness of the generalized Borda Count ranking, and use synthetic and real data sets to verify the efficiency of our computational method.        2014
3. Rough Sets, Kernel Set, and Spatiotemporal Outlier Detection Abstract—Nowadays, the high availability of data gathered from wireless sensor networks and telecommunication systems has drawn the attention of researchers on the problem of extracting knowledge from spatiotemporal data. Detecting outliers which are grossly different from or inconsistent with the remaining spatiotemporal data set is a major challenge in real-world knowledge discovery and data mining applications. In this paper, we deal with the outlier detection problem in spatiotemporal data and describe a rough set approach that finds the top outliers in an unlabeled spatiotemporal data set. The proposed method, called Rough Outlier Set Extraction (ROSE), relies on a rough set theoretic representation of the outlier set using the rough set approximations, i.e., lower and upper approximations. We have also introduced a new set, named Kernel Set, that is a subset of the original data set, which is able to describe the original data set both in terms of data structure and of obtained results. Experimental results on  real-world data sets demonstrate the superiority of ROSE, both in terms of some quantitative indices and outliers detected, over those obtained by various rough fuzzy clustering algorithms and by the state-of-the-art outlier detection methods. It is also demonstrated that the kernel set is able to detect the same outliers set but with less computational time.        2014
4. Discovering Temporal Change Patterns in the Presence of Taxonomies Frequent items mining is a widely exploratory technique that focuses on discovering recurrent correlations among data. The steadfast evolution of markets and business environments prompts the need of data mining  algorithms to discover significant correlation changes in order to reactively suit product and service provision to customer needs. Change mining, in the context of frequent item sets, focuses on detecting and reporting significant changes in the set of mined item sets from one time period to another. The discovery of frequent generalized item sets, i.e., item sets that 1) frequently occur in the source data, and 2) provide a high-level abstraction of the mined knowledge, issues new challenges in the analysis of item sets that become rare, and thus are no longer extracted, from a certain point. This paper proposes a novel kind of dynamic pattern, namely the History Generalized Pattern (HiGen), that represents the evolution of an item set in consecutive time periods, by reporting the information about its frequent generalizations characterized by minimal redundancy (i.e., minimum level of abstraction) in case it becomes infrequent in a certain time period. To address HiGen mining, it proposes HiGen Miner, an algorithm that focuses on avoiding item set mining followed by post processing by exploiting a support-driven item set generalization approach. To focus the attention on the minimally redundant frequent generalizations and thus reduce the amount of the generated patterns, the discovery of a smart subset of HiGens, namely the Non-redundant HiGens, is addressed as well. Experiments performed on both real and synthetic datasets show the efficiency and the effectiveness of the proposed approach as well as its usefulness in a

real application context.

       2013
5. Information-Theoretic Outlier Detection for Large-Scale Categorical Data

 

Outlier detection can usually be considered as a pre-processing step for locating, in a data set, those objects that do not conform to well-defined notions of expected behavior. It is very

important in data mining for discovering novel or rare events, anomalies, vicious actions, exceptional phenomena, etc. We are investigating outlier detection for categorical data sets. This problem is especially challenging because of the difficulty of defining a meaningful similarity measure for categorical data. In this paper, we propose a formal definition of outliers and an optimization model of outlier detection, via a new concept of holo entropy that takes both entropy and total correlation into consideration. Based on this model, we define a function for the outlier factor of an object which is solely determined by the object itself and can be updated efficiently. We propose two practical 1

-parameter outlier detection methods, named ITB-SS and ITB-SP, which require no user-defined parameters for deciding whether an object is an outlier. Users need only provide the number of outliers they want to detect. Experimental results show that ITB-SS and ITB-SP are more effective and efficient than mainstream methods and can be used to deal with both large and high-dimensional data sets where existing algorithms fail.

       2013
6. Robust Module-Based Data Management

 

The current trend for building an ontology-based data management system (DMS) is to capitalize on efforts made to design a preexisting well-established DMS (a reference system). The method amounts to extracting from the reference DMS a piece of schema relevant to the new application needs-a module

-, possibly personalizing it with extra constraints w.r.t. the application under construction, and then managing a data set using the resulting schema. In this paper, we extend the existing definitions of modules and we introduce novel properties of robustness that provide means for checking easily that a robust module-based DMS evolves safely w.r.t. both the schema and the data of the reference DMS. We carry out our investigations in the setting of description logics which underlie modern ontology languages, like RDFS, OWL, and OWL2 from W3C. Notably, we focus on the DL-liteA dialect of the DL-lite family, which encompasses the foundations of the QL profile of OWL2 (i.e., DL-liteR): the W3C recommendation for efficiently managing large data sets.

       2013
7. Protecting Sensitive Labels in Social Network Data Anonymization

 

Privacy is one of the major concerns when publishing or sharing social network data for social science research and business analysis. Recently, researchers have developed privacy models similar to k-anonymity to prevent node reidentification through structure information. However, even when these privacy models are enforced, an attacker may still be able to infer one’s private information if a group of nodes largely share the same sensitive labels (i.e., attributes). In other words, the label-node relationship is not well protected by pure structure  anonymization methods. Furthermore, existing approaches, which rely on edge editing or node clustering, may significantly alter key graph properties. In this paper, we define a k-degree-l-diversity anonymity model that considers the protection of structural information as well as sensitive labels of individuals. We further propose a novel anonymization methodology based on adding noise nodes. We develop a new algorithm by adding noise nodes into the original graph with the consideration of introducing the least distortion to graph properties. Most importantly, we provide a rigorous analysis of the theoretical bounds on the number of noise nodes added and their impacts on an important graph property. We conduct extensive experiments to evaluate the effectiveness of the proposed technique        2013
8. TrustedDB: A Trusted Hardware Based Database with Privacy and Data Confidentiality Traditionally, as soon as confidentiality becomes a concern, data is encrypted before outsourcing to a service provider. Any software-based cryptographic constructs then deployed, for server-side query processing on the encrypted data, inherently limit query expressiveness. Here, we introduce TrustedDB, an outsourced database prototype that allows clients to execute SQL queries with privacy and under regulatory compliance constraints by leveraging server-hosted, tamper-proof trusted hardware in critical query processing stages, thereby removing any limitations on the type of supported queries. Despite the cost overhead and performance limitations of trusted hardware, we show that the costs per query are orders of magnitude lower than any (existing or) potential future software-only mechanisms. TrustedDB is built and runs on actual hardware, and its performance and costs are evaluated here.        2013

 

Copyright © 2025 · News Pro Theme on Genesis Framework · WordPress · Log in