Software Similarity and Classification


Book Description

Software similarity and classification is an emerging topic with wide applications. It is applicable to the areas of malware detection, software theft detection, plagiarism detection, and software clone detection. Extracting program features, processing those features into suitable representations, and constructing distance metrics to define similarity and dissimilarity are the key methods to identify software variants, clones, derivatives, and classes of software. Software Similarity and Classification reviews the literature of those core concepts, in addition to relevant literature in each application and demonstrates that considering these applied problems as a similarity and classification problem enables techniques to be shared between areas. Additionally, the authors present in-depth case studies using the software similarity and classification techniques developed throughout the book.




Intelligent Systems in Operations: Methods, Models and Applications in the Supply Chain


Book Description

"This book provides knowledge and insights on present and future AI applications in Operations Management presenting tools and decisions in terms of theoretical and empirical models, methods and proposed applications"--Provided by publisher.




Software Technologies


Book Description

This book constitutes the refereed proceedings of the 16th International Conference on Software Technologies, ICSOFT 2021, Virtual Event, July 6–8, 2021. The conference was held virtually due to the COVID-19 crisis. The 10 full papers included in this book were carefully reviewed and selected from 117 submissions.




Tools and Methods of Program Analysis


Book Description

This book constitutes the refereed proceedings of the 5th International Conference on Tools and Methods for Program Analysis, TMPA 2019, held in Tbilisi, Georgia, in November 2019. The 14 revised full papers and 2 revised short papers presented together with one keynote paper were carefully reviewed and selected from 41 submissions. The papers deal with topics such as software test automation, static program analysis, verification, dynamic methods of program analysis, testing and analysis of parallel and distributed systems, testing and analysis of high-load and high-availability systems, analysis and verification of hardware and software systems, methods of building quality software, tools for software analysis, testing and verification.




SUPERVISED LEARNING ALGORITHMS CLASSIFICATION AND REGRESSION ALGORITHMS


Book Description

The branch of computer science known as machine learning is one of the subfields that is increasing at one of the fastest rates now and has various potential applications. The technique of automatically locating meaningful patterns in vast volumes of data is referred to as pattern recognition. It is possible to provide computer programs the ability to learn and adapt in response to changes in their surroundings via the use of tools for machine learning. As a consequence of machine learning being one of the most essential components of information technology, it has therefore become a highly vital, though not always visible, component of our day-to-day life. As the amount of data that is becoming available continues to expand at an exponential pace, there is good reason to believe that intelligent data analysis will become even more common as a critical component for the advancement of technological innovation. This is because there is solid grounds to believe that this will occur. Despite the fact that data mining is one of the most significant applications for machine learning (ML), there are other uses as well. People are prone to make mistakes while doing studies or even when seeking to uncover linkages between a lot of distinct aspects. This is especially true when the analyses include a large number of components. Data Mining and Machine Learning are like Siamese twins; from each of them, one may get a variety of distinct insights by using the right learning methodologies. As a direct result of the development of smart and nanotechnology, which enhanced people's excitement in discovering hidden patterns in data in order to extract value, a great deal of progress has been achieved in the field of data mining and machine learning. These advancements have been very beneficial. There are a number of probable explanations for this phenomenon, one of which is that people are currently more inquisitive than ever before about identifying hidden patterns in data. As the fields of statistics, machine learning, information retrieval, and computers have grown increasingly interconnected, we have seen an increase in the led to the development of a robust field that is built on a solid mathematical basis and is equipped with extremely powerful tools. This field is known as information theory and statistics. The anticipated outcomes of the many different machine learning algorithms are culled together into a taxonomy that is used to classify the many different machine learning algorithms. The method of supervised learning may be used to produce a function that generates a mapping between inputs and desired outputs. The production of previously unimaginable quantities of data has led to a rise in the degree of complexity shown across a variety of machine learning strategies. Because of this, the use of a great number of methods for both supervised and unsupervised machine learning has become obligatory. Because the objective of many classification challenges is to train the computer to learn a classification system that we are already familiar with, supervised learning is often used in order to find solutions to problems of this kind. The goal of unearthing the accessibility hidden within large amounts of data is well suited for the use of machine learning. The ability of machine learning to derive meaning from vast quantities of data derived from a variety of sources is one of its most alluring prospects. Because data drives machine learning and it works on a large scale, this goal will be achieved by decreasing the amount of dependence that is put on individual tracks. Machine learning functions on data. Machine learning is best suited towards the complexity of managing through many data sources, the huge diversity of variables, and the amount of data involved, since ML thrives on larger datasets. This is because machine learning is ideally suited towards managing via multiple data sources. This is possible as a result of the capacity of machine learning to process ever-increasing volumes of data. The more data that is introduced into a framework for machine learning, the more it will be able to be trained, and the more the outcomes will entail a better quality of insights. Because it is not bound by the limitations of individual level thinking and study, ML is intelligent enough to unearth and present patterns that are hidden in the data.




Citation-based Plagiarism Detection


Book Description

Plagiarism is a problem with far-reaching consequences for the sciences. However, even today’s best software-based systems can only reliably identify copy & paste plagiarism. Disguised plagiarism forms, including paraphrased text, cross-language plagiarism, as well as structural and idea plagiarism often remain undetected. This weakness of current systems results in a large percentage of scientific plagiarism going undetected. Bela Gipp provides an overview of the state-of-the art in plagiarism detection and an analysis of why these approaches fail to detect disguised plagiarism forms. The author proposes Citation-based Plagiarism Detection to address this shortcoming. Unlike character-based approaches, this approach does not rely on text comparisons alone, but analyzes citation patterns within documents to form a language-independent "semantic fingerprint" for similarity assessment. The practicability of Citation-based Plagiarism Detection was proven by its capability to identify so-far non-machine detectable plagiarism in scientific publications.




NASA Technical Note


Book Description




Advances in Computer and Information Sciences and Engineering


Book Description

Advances in Computer and Information Sciences and Engineering includes a set of rigorously reviewed world-class manuscripts addressing and detailing state-of-the-art research projects in the areas of Computer Science, Software Engineering, Computer Engineering, and Systems Engineering and Sciences. Advances in Computer and Information Sciences and Engineering includes selected papers from the conference proceedings of the International Conference on Systems, Computing Sciences and Software Engineering (SCSS 2007) which was part of the International Joint Conferences on Computer, Information and Systems Sciences and Engineering (CISSE 2007).




Advanced Informatics for Computing Research


Book Description

This book constitutes the refereed proceedings of the First International Conference on Advanced Informatics for Computing Research , ICAICR 2017, held in Jalandhar, India, in March 2017. The 32 revised full papers presented were carefully reviewed and selected from 312 submissions. The papers are organized in topical sections on computing methodologies, information systems, security and privacy, network services.




Advances on P2P, Parallel, Grid, Cloud and Internet Computing


Book Description

This book presents the latest research findings, innovative research results, methods and development techniques related to P2P, grid, cloud and Internet computing from both theoretical and practical perspectives. It also reveals the synergies among such large-scale computing paradigms. P2P, grid, cloud and Internet computing technologies have rapidly become established as breakthrough paradigms for solving complex problems by enabling aggregation and sharing of an increasing variety of distributed computational resources at large scale. Grid computing originated as a paradigm for high-performance computing, as an alternative to expensive supercomputers through different forms of large-scale distributed computing. P2P computing emerged as a new paradigm after client–server and web-based computing and has proved useful in the development of social networking, B2B (business to business), B2C (business to consumer), B2G (business to government), and B2E (business to employee). Cloud computing has been defined as a “computing paradigm where the boundaries of computing are determined by economic rationale rather than technical limits,” and it has fast become a computing paradigm with applicability and adoption in all application domains and which provides utility computing at a large scale. Lastly, Internet computing is the basis of any large-scale distributed computing paradigms; it has developed into a vast area of flourishing fields with enormous impact on today’s information societies, and serving as a universal platform comprising a large variety of computing forms such as grid, P2P, cloud and mobile computing.