We Provide computer science,Information Technology and Computer Engineering related
projects for ME,MTech and MS students. You can download your project from our readymade
project list specifically collected for Master of Engineering(ME),Master Of Technology(MTech)
and Master of Science(MS).It includes unexplored unique topics that we are collected
form different online resources. We provide this Master degree project for your
guide to complete your thesis or dissertation of Master of Engineering(ME),Master
of Technology(MTech) and Master of Science(MS). We have also uploaded project available
in different technologies like Java, C, C++, Dot Net, Android, Php, Ns2 etc. to
cater your project needs.
ARP Spoofing:A Comparative Study for Education Purposes
ARP spoofing attack, one of the most important security topics, is usually taught
in courses such as Intrusion Detection in Local Area Networks (LANs). In such a
course, hands-on labs are very important as they facilitate students’ learning on
how to detect ARP spoofing using various types of security solutions, such as intrusion
detection and prevention systems (IDS/IPS). The preparation of these hands-on labs
are usually the task of Security Instructors who are required to select and use
efficient security solutions for their hands-on experiments; the problem that presents
itself is that most of these security instructors lack the sufficient hands-on experience
and skills. For this reason and because of the diversity of the available security
solutions, the security instructors are having much difficulty when selecting the
adequate security solutions for their hands-on labs. This paper is a comparative
study for educational purpose. It provides analysis based on practical experiments
carried out on a number of security solutions regarding their ability to detect
ARP spoofing. Our analysis provides means for security instructors to evaluate and
select the appropriate security solutions for their hands-on labs. In addition,
we clearly show that ARP spoofing has not been given enough attention by most tested
security solutions, even though this attack presents a serious threat, is very harmful
and more dangerously it is easy to conduct. As a solution, we propose the requirements
for an ideal algorithm that can be used by security solutions to detect effectively
any ARP spoofing attack.
Keywords:- ARP spoofing, ARP spoofing detection, Denial of Service (DoS)
|
Classification and Clustering-Using Intelligent Techniques
Analysis and interpretation of DNA Microarray data is a fundamental task in bioinformatics.
Feature Extraction plays a critical role in better performance of the classifier.
We address the dimension reduction of DNA features in which relevant features are
extracted among thousands of irrelevant ones through dimensionality reduction. This
enhances the speed and accuracy of the classifiers. Principal Component Analysis
(PCA) is a technique used for feature extraction which helps to retrieve intrinsic
information from high dimensional data in eigen spaces to solve the curse of dimensionality
problem. The curse of dimensionality means n >> m, where n is a large number of
features and m is a small number of samples (may be too less). Neural Networks (NN)
and Support Vector Machine (SVM) are implemented and their performances are measured
in terms of predictive accuracy, specificity, and sensitivity. First, we implement
PCA for significant feature extraction and then FFNN trained using Backpropagation
(BP) and SVM are implemented on the reduced feature set. Next, we propose a Multiobjective
Genetic Algorithm-based fuzzy clustering technique using real coded encoding of
cluster centers for clustering and clas- si cation. This technique is implemented
on microarray cancer data to select training data using multiobjective genetic algorithm
with non-dominated sorting (MOGA-NSGA-II). The two objective functions for this
multiobjective techniques are optimization of cluster compactness as well as separation
simultaneously. This approach identifies the solution i.e. the individual chromosome
which gives the optimal value of the compactness and separation. Then we find high
con dence points for these non-dominated set using a fuzzy voting technique. Support
Vector Machine (SVM) classifier is further trained by the selected training points
which have high confidence value. Then remaining points are classified by trained
SVM classifier. Finally, the four clustering label vectors through majority vot-
ing ensemble are combined, i.e., each point is assigned a class label that obtains
he maximum number of votes among the four clustering solutions. The performance
of the proposed MOGA-SVM, classification and clustering method has been compared
to MOGA-BP, SVM, BP. The performance are measured in terms of Silhoutte Index, ARI
Index respectively. The experiment were carried on three public domain cancer data
sets, viz., Ovarian, Colon and Leukemia cancer data to establish its superiority.
Keywords:- Cancer Classification; Feature Reduction; Multiobjective genetic al-
gorithm; Neural Network; Pareto-optimality; Principal components; Support Vec- tor
Machine(SVM)
|
Group Signature Scheme Resistant against Colluding Attack
Group signature is an extension of digital signature, which allows a group member
to sign anonymously a document on behalf of the group. Any client can verify the
authenticity of the document by using the public parameters of the group. The identity
of the group member cannot be revealed from the group signature. In case of a legal
dispute, an authorized group member can disclose the identity of the group member
from the signed document. Group signature can have wide application to corporate
world, banks, and e-commerce applications. In this thesis, we designed a group signature
protocol based upon hard computa- tional assumptions such as, Discrete Logarithm
Problem (DLP), Integer Factor- ization Problem (IFP), and Computational Die Hellmann
(CDH) problem. The proposed scheme is proved to be resistant against colluding attack.
Moreover, the group signature remains valid, if some members leave the group or
some new mem- bers join the group. Full traceability feature is con rmed in the
proposed scheme. The scheme can have wide applications in real life scenarios such
as e-banking, e-voting, and e-commerce applications.
Keywords:- anonymity; colluding attack; discrete logarithm; group signature; unforgeability
|
Image Deblurring in Presence of Gaussian and Impulsive Noise
Image restoration is an essential and unavoidable preprocessing operation for many
security applications like biometric security, video surveillance, object tracking,
image data communication etc. Images are generally degraded due to faulty sensor,
channel transmission error, camera mis-focus, atmospheric turbulence, relative motion
between camera and object etc. Such conditions are inevitable while capturing a
scene through camera. Restoration of such images is highly essential for further
image processing and other tasks.
Keywords:- Image restoration, Impulsive noise, Gaussian noise, Motion blur, Out-of-focus
blur, Regularization, Convex minimization.
|
Helmholtz Principle-Based Keyword Extraction
In today’s world of evolving technology, everybody wishes to accomplish tasks in
least time. As information available online is perpetuating every day, it becomes
very dicult to summarize any more than 100 documents in acceptable time. Thus,
”text summarization” is a challenging problem in the area of Natural Language Processing
(NLP) especially in the context of global languages. In this thesis, we survey taxonomy
of text summarization from di erent aspects. It briefly explains di erent approaches
to summarization and the evaluation parameters. Also presented are a thorough details
and facts about more than fifty automatic text summarization systems to ease the
job of researchers and serve as a short encyclopedia for the investigated systems.
Keyword extraction methods plays vital role in text mining and document processing.
Keywords represent essential content of a document. Text mining applications take
the advantage of keywords for processing documents. A quality Keyword is a word
that represents the exact content of the text subsetly. It is very dicult to process
large number of documents to get high quality keywords in acceptable time. This
thesis gives a comparison between the most popular keyword extractions method, tf-idf
and the proposed method that is based on Helmholtz Principle. Helmholtz Principle
is based on the ideas from image processing and derived from the Gestalt theory
of human perception. We also investigate the run time to extract the keywords by
both the methods. Experimental results show that keyword extraction method based
on Helmholtz Principle outperformancetf-idf.
Keywords:- Keywords: Text Mining, Text Summarization, Stemming, Helmholtz Peinciple,
Information Retrieval, Keyword Extraction, Term Frequency - Inverse Document Frequency.
|
Load Balancing in MANET : Alleviating the center node
Load balancing is an essential requirement of any multi-hop wireless network. A
wireless routing protocol is accessed on its ability to distribute traffic over
the network nodes and a good routing protocol achieves this without introducing
un- acceptable delay. The most obvious benefit is manifested in increasing the life
of a battery operated node which can eventually increase the longevity of the entire
network. In the endeavor of finding the shortest distance between any two nodes
to transmit data fast the center nodes become the famous picks. The centrally located
nodes connect many subnetworks and serve as gateways to some subnetworks that become
partitioned from the rest of the network in its absence. Thus, the lifetime of the
center nodes become a bottleneck for connectivity of a subnetwork prior to its partition
from the rest of the network. An unbiased load can cause congestion in the network
which impacts the overall throughput, packet delivery ratio and the average end
to end delay. In, this thesis we have mitigated the unbiased load distribution on
centrally located nodes by pushing traffic further to the peripheral nodes without
compromising the average end to end delay for a greater network longevity and performances.
We proposed a novel routing metric , load and a minimization criterion to decide
a path that involves nodes with less load burden on them. The simulations of the
proposed mechanism run on NS-2.34 for 16 and 50 nodes have revealed an average 2.26%
reduction of load on the center node in comparison with AOMDV.
Keywords:- Routing Protocol,Wireless Routing Algorithm,Improving performance of
MANET
|
Software Defect Prediction Based on Classification Rule Mining
There has been rapid growth of software development. Due to various causes, the
software comes with many defects. In Software development process, testing of software
is the main phase which reduces the defects of the software. If a developer or a
tester can predict the software defects properly then, it reduces the cost, time
and effort. In this paper, we show a comparative analysis of software defect prediction
based on classification rule mining. We propose a scheme for this process and we
choose different classication algorithms. Showing the comparison of predictions
in software defects analysis. This evaluation analyzes the prediction performance
of competing learning schemes for given historical data sets(NASA MDP Data Set).
The result of this scheme evaluation shows that we have to choose different classifier
rule for different data set.
Keywords:- Software defect prediction, classification Algorithm, Cofusion matrix.
|
Evaluation of Software Understandability Using Software Metrics
Understandability is one of the important characteristics of software quality, because
it may influence the maintainability of the software. Cost and reuse of the software
is also affected by understandability. In order to maintain the software, the programmers
need to understand the source code. The understandability of the source code depends
upon the psychological complexity of the software, and it requires cognitive abilities
to understand the source code. The understandability of source code is get effected
by so many factors, here we have taken different factors in an integrated view.
In this we have chosen rough set approach to calculate the understandability based
on outlier detection. Generally the outlier is having an abnormal behavior, here
we have taken that project has may be easily understandable or difficult to understand.
Here we have taken few factors, which affect understandability, an brings forward
an integrated view to determine understandability.
Keywords:- Understandability, Rough set, Outlier, Spatial Complexity.
|
Upgrading Shortest Paths in Networks
We introduce the Upgrading Shortest Paths Problem, a new combinatorial problem for
improving network connectivity with a wide range of applications from multicast
communication to wildlife habitat conservation. We define the problem in terms of
a network with node delays and a set of node upgrade actions, each associated with
a cost and an upgraded (reduced) node delay. The goal is to choose a set of upgrade
actions to minimize the shortest delay paths between demand pairs of terminals in
the network, subject to a budget constraint. We show that this problem is NP-hard.
We describe and test two greedy algorithms against an exact algorithm on synthetic
data and on a real-world instance from wildlife habitat conservation. While the
greedy algorithms can do arbitrarily poorly in the worst case, they perform fairly
well in practice. For most of the instances, taking the better of the two greedy
solutions accomplishes within 5% of optimal on our benchmarks.
Keywords:- Shortest Path Problem,improving Network connectivity,demand pairs.
|
Improved Modified Condition/ Decision Coverage using Code Transformation Techniques
ModiFied Condition / Decision Coverage (MC / DC) is a thesis/Dissertation Submitted
For Master in Engineering.ModiFied Condition / Decision Coverage (MC / DC) is a
white box testing criteria aiming to prove that all conditions involved in a predicate
can in uence the predicate value in the desired way. In regulated domains such as
aerospace and safety critical domains, software quality assurance is subjected to
strict regulations such as the DO-178B standard. Though MC/DC is a standard coverage
criterion, existing automated test data genera- tion approaches like CONCOLIC testing
do not support MC/DC. To address this issue we present an automated approach to
generate test data that helps to achieve an increase in MC/DC coverage of a program
under test. We use code transformation techniques for transforming program. This
transformed program is inserted into the CREST TOOL. It drives CREST TOOL to generate
test suite and increase the MC/DC coverage. Our tech- nique helps to achieve a signi
cant increase in MC/DC coverage as compared to traditional CONCOLIC testings. Our
experimental results show that the proposed approach helps to achieve on the average
approximately 20.194 % for Program Code Transformer(PCT) and 25.447 % for Exclusive-
Nor Code Transformer. The average time taken for seventeen programs is 6.89950 seconds.
Keywords:- CONCOLIC testing, Code transformation techniques, MC/DC, Coverage Analyser,ME
Thesis,Master Dissertation
|
Methods for Redescription Mining
In scientific investigations data oftentimes have different nature. For instance,
they might originate from distinct sources or be cast over separate terminologies.
In order to gain insight into the phenomenon of interest, a natural task is to identify
the correspondences that exist between these different aspects. This is the motivating
idea of redescription mining, the data analysis task studied in this thesis. Redescription
mining aims to find distinct common characterizations of the same objects and, vice
versa, to identify sets of objects that admit multiple shared descriptions. A practical
example in biology consists in finding geographical areas that admit two characterizations,
one in terms of their climatic profile and one in terms of the occupying species.
Discovering such redescriptions can contribute to better our understanding of the
influence of climate over species distribution. Besides biology, applications of
redescription mining can be envisaged in medicine or sociology, among other elds.
Previously, redescription mining was restricted to propositional queries over Boolean
attributes. However, many conditions, like aforementioned climate, cannot be expressed
naturally in this limited formalism. In this thesis, we consider more general query
languages and propose algorithms to find the corresponding redescriptions, making
the task relevant to a broader range of domains and problems. Specifically, we start
by extending redescription mining to non-Boolean attributes. In other words, we
propose an algorithm to handle nominal and real-valued attributes natively. We then
extend redescription mining to the relational setting, where the aim is to find
corresponding connection patterns that relate almost the same object tuples in a
network.
Keywords:- Data Mining Thesis,ME Thesis,Master Dissertation,Redescription mining
|
Enabling Multipath and Multicast Data Transmission in Legacy and Future Internet
The quickly growing community of Internet users is requesting multiple applications
and services. At the same time the structure of the network is changing. From the
performance point of view, there is a tight interplay between the application and
the network design. The network must be constructed to provide an adequate performance
of the target application. In this thesis we consider how to improve the quality
of users’ experience concentrating on two popular and resource-consuming applications:
bulk data transfer and real-time video streaming. We share our view on the techniques
which enable feasibility and deployability of the network functionality leading
to unquestionable performance improvement for the corresponding applications. Modern
mobile devices, equipped with several network interfaces, as well as multihomed
residential Internet hosts are capable of maintaining multiple simultaneous attachments
to the network. We propose to enable simultaneous multipath data transmission in
order to increase throughput and speed up such bandwidth-demanding applications
as, for example, file download. We design an extension for Host Identity Protocol
(mHIP), and propose a multipath data scheduling solution on a wedge layer between
IP and transport, which effectively distributes packets from a TCP connection over
available paths. We support our protocol with a congestion control scheme iii iv
and prove its ability to compete in a friendly manner against the legacy network
protocols. Moreover, applying game-theoretic analytical modelling we investigate
how the multihomed HIP multipath-enabled hosts coexist in the shared network. The
number of real-time applications grows quickly.
Efficient and reliable transport of multimedia content is a critical issue of today’s
IP network design. In this thesis we solve scalability issues of the multicast dissemination
trees controlled by the hybrid error correction. We propose a scalable multicast
architecture for potentially large overlay networks. Our techniques address suboptimality
of the adaptive hybrid error correction (AHEC) scheme in the multicast scenarios.
A hierarchical multi-stage multicast tree topology is constructed in order to improve
the performance of AHEC and guarantee QoS for the multicast clients. We choose an
evolutionary networking approach that has the potential to lower the required resources
for multimedia applications by utilizing the error-correction domain separation
paradigm in combination with selective insertion of the supplementary data from
parallel networks, when the corresponding content is available.
Keywords:- Multipath Data Transmission,Multicast Network Thesis,ME Thesis,Master
Dissertation,Computer Network
|
Data fusion and matching by maximizing statistical dependencies
Multi-view learning is a task of learning from multiple data sources where each
source represents a different view of the same phenomenon. Typ- ical examples include
multimodal information retrieval and classification of genes by combining heterogeneous
genomic data. Multi-view learning methods can be motivated by two interrelated lines
of thoughts: if single view is not sufficient for the learning task, other views
can complement the information. Secondly, learning by searching for an agreement
between views may generalize better than learning from a single view. In this thesis,
novel methods for unsupervised multi-view learning are proposed. Multi-view learning
methods, in general, work by searching for an agree- ment between views. However,
defining an agreement is not straightforward in an unsupervised learning task. In
this thesis, statistical dependency is used to define an agreement between the views.
Assuming that the shared information between the views is more interesting, statistical
dependency is used to find the shared information. Based on this principle, a fast
linear preprocessing method that performs data fusion during exploratory data analysis
is introduced. Also, a novel evaluation approach based on the dependency between
views to compare vector representations for bilingual corpora is introduced.
Keywords:- Data Fusion,Multi-View Learning Thesis,ME Thesis,Master Dissertation
|
INDEXING OF LARGE BIOMETRIC DATABASE
The word "biometrics" is derived from the Greek words 'bios' and 'metric' which
means life and measurement respectively. This directly translates into "life measurement".
Biometrics is the automated recognition of individuals based on their behavioral
and biological characteristics. Biometric features are information extracted from
biometric samples which can be used for comparison with a biometric reference. Biometrics
comprises methods for uniquely recognizing humans based upon one or more intrinsic
physical or behavioral traits. In computer science, in particular, biometrics is
used as a form of identity access management and access control. It is also used
to identify individuals in groups that are under surveillance. Biometrics has fast
emerged as a promising technology for authentication and has already found place
in most hi-tech security areas. An efficient clustering technique has been proposed
for partitioning large biometric database during identification. The system has
been tested using bin-miss rate as a performance parameter. As we are still getting
a higher bin-miss rate, so this work is based on devising an indexing strategy for
identification of large biometric database and with greater accuracy. This technique
is based on the modified B+ tree which reduces the disk accesses. It decreases the
data retrieval time and also possible error rates. The indexing technique is used
to declare a person‟s identity with lesser number of comparisons rather than searching
the entire database. The response time deteriorates, as well as the accuracy of
the system degrades as the size of the database increases. Hence for larger applications,
the need to reduce the database to a smaller fraction arises to achieve both higher
speeds and improved accuracy. The main purpose of indexing is to retrieve a small
portion of the database for searching the query. Since applying some traditional
clustering schemes does not yield satisfactory results, we go for an indexing strategy
based on tree data structures. Index is used to look-up, input and delete data in
an ordered manner. Speed and efficiency are the main goals in the different types
of indexing. Speed and efficiency include factors like access time, insertion time,
deletion time, and space overhead. The main aim is to perform indexing of a database
using different trees beginning with Binary Search tree followed by B tree before
proceeding to its variations, B+ tree and Modified B+ tree, and subsequently determine
their performance based on their respective execution times.
Keywords:-Biometric Data,Tree DataStructure,Biometric Database,Thesis,ME Thesis,Master
Dissertation
|
Intrusion Detection Using Self-Training Support Vector Machines
Intrusion is broadly defined as a successful attack on a network. The definition
of attack itself is quite ambiguous and there exists various de nitions of it. With
the advent of Internet age and the tremendous increase in the computational resources
available to an average user, the security risk of each and every computer has grown
exponentially. Intrusion Detection System (IDS) is a software tool used to detect
unauthorized access to a computer system or network. It is a dynamic monitoring
entity that complements the static monitoring abilities of a rewall. Data Mining
techniques provide efficient methods for the development of IDS. The idea behind
using data mining techniques is that they can automate the process of creating traffic
models from some reference data and thereby eliminate the need of laborious manual
intervention. Such systems are capable of detecting not only known attacks but also
their variations. Existing IDS technologies, on the basis of detection methodology
are broadly clas- sified as Misuse or Signature Based Detection and Anomaly Detection
Based System. The idea behind misuse detection consists of comparing network traffic
against a Model describing known intrusion. The anomaly detection method is based
on the analysis of the pro les that represent normal traffic behavior. Semi-Supervised
systems for anomaly detection would reduce the demands of the training process by
reducing the requirement of training labeled data. A Self Training Support Vector
Machine based detection algorithm is presented in this thesis. In the past, Self-Training
of SVM has been successfully used for reducing the size of labeled training set
in other domains. A similar method was implemented and results of the simulation
performed on the KDD Cup 99 dataset for intrusion detection show a reduction of
upto 90% in the size of labeled training set required as compared to the supervised
learning techniques.
Keywords:-Intrusion Detection System,IDS,Network Security Thesis,Vector Learning,ME
Thesis,Master Dissertation
|
Automatic Detection of Fake Profiles in Online Social Networks
In the present generation, the social life of everyone has become associated with
the online social networks. These sites have made a drastic change in the way we
pursue our social life. Making friends and keeping in contact with them and their
updates has become easier. But with their rapid growth, many problems like fake
profiles, online impersonation have also grown. There are no feasible solution exist
to control these problems. In this project, we came up with a framework with which
automatic detection of fake profiles is possible and is efficient. This framework
uses classification techniques like Support Vector Machine, Nave Bayes and Decision
trees to classify the profiles into fake or genuine classes. As, this is an automatic
detection method, it can be applied easily by online social networks which has millions
of profile whose profiles can not be examined manually.
Keywords:- Social Networking,Fake Profile Detection,Vector Learning,ME Thesis,Master
Dissertation
|
Query-Time Optimization Techniques for Structured Queries in Information Retrieval
The use of information retrieval (IR) systems is evolving towards larger, more complicated
queries. Both the IR industrial and research communities have generated significant
evidence indicating that in order to continue improving retrieval effectiveness,
increases in retrieval model complexity may be unavoidable. From an operational
perspective, this translates into an increasing computational cost to generate the
final ranked list in response to a query. Therefore we encounter an increasing tension
in the trade-o between retrieval effectiveness (quality of result list) and efficiency
(the speed at which the list is generated). This tension creates a strong need for
optimization techniques to improve the efficiency of ranking with respect to these
more complex retrieval models.
This thesis presents three new optimization techniques designed to deal with different
aspects of structured queries. The first technique involves manipulation of interpolated
subqueries, a common structure found across a large number of retrieval models today.
We then develop an alternative scoring formulation to make retrieval models more
responsive to dynamic pruning techniques. The last technique is delayed execution,
which focuses on the class of queries that utilize term dependencies and term conjunction
operations. In each case, we empirically show that these optimizations can significantly
improve query processing efficiency without negatively impacting retrieval effectiveness.
Keywords:- information retrieval (IR) systems,Query Optimization,Query processing,ME
Thesis,Master Dissertation
|
The Security and Privacy Implications of Energy- Proportional Computing
The parallel trends of greater energy-efficiency and more aggressive power management
are yielding computers that inch closer to energy-proportional computing with every
generation. Energy-proportional computing, in which power consumption scales closely
with workload, has unintended side effects for security and privacy. Saving energy
is an unqualified boon for computer operators, but it is becoming easier to identify
computing activities by observing power consumption because an energy-proportional
computer reveals more about its workload. This thesis demonstrates the potential
for system-level power analysis—the inference of a computers internal states based
on power observation at the “plug.” It also examines which hardware components and
software workloads have the greatest impact on information leakage. This thesis
identifies the potential for privacy violations by demonstrating that a malicious
party could identify which webpage from a given corpus a user is viewing with greater
than 99% accuracy. It also identifies constructive applications for power analysis,
evaluating its use as an anomaly detection mechanism for embedded devices with greater
than 94% accuracy for each device tested. Finally, this thesis includes modeling
work that correlates AC and DC power consumption to pinpoint which components contribute
most to information leakage and analyzes software workloads to identify which classes
of work lead to the most information leakage. Understanding the security and privacy
risks and opportunities that come with energy-proportional computing will allow
future systems to either apply system-level power analysis fruitfully or thwart
its malicious application.
Keywords:- Social Networking,Fake Profile Detection,Vector Learning,ME Thesis,Master
Dissertation
|