In this book, the following three approaches to data analysis are presented: - Test Theory, founded by Sergei V. Yablonskii (1924-1998); the first publications appeared in 1955 and 1958, - Rough Sets, founded by Zdzisław I. Pawlak (1926-2006); the first publications appeared in 1981 and 1982, - Logical Analysis of Data, founded by Peter L. Hammer (1936-2006); the first publications appeared in 1986 and 1988. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected. - Logical Analysis of Data, founded by Peter L. Hammer (1936-2006); the first publications appeared in 1986 and 1988. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected.
In this book, the concept of a hypothesis about the values of all attributes is added to the standard decision tree model, considered, in particular, in test theory and rough set theory. This extension allows us to use the analog of equivalence queries from exact learning and explore decision trees that are based on various combinations of attributes, hypotheses, and proper hypotheses (analog of proper equivalence queries). The two main goals of this book are (i) to provide tools for the experimental and theoretical study of decision trees with hypotheses and (ii) to compare these decision trees with conventional decision trees that use only queries, each based on a single attribute. Both experimental and theoretical results show that decision trees with hypotheses can have less complexity than conventional decision trees. These results open up some prospects for using decision trees with hypotheses as a means of knowledge representation and algorithms for computing Boolean functions. The obtained theoretical results and tools for studying decision trees with hypotheses are useful for researchers using decision trees and rules in data analysis. This book can also be used as the basis for graduate courses.
The results presented here (including the assessment of a new tool – inhibitory trees) offer valuable tools for researchers in the areas of data mining, knowledge discovery, and machine learning, especially those whose work involves decision tables with many-valued decisions. The authors consider various examples of problems and corresponding decision tables with many-valued decisions, discuss the difference between decision and inhibitory trees and rules, and develop tools for their analysis and design. Applications include the study of totally optimal (optimal in relation to a number of criteria simultaneously) decision and inhibitory trees and rules; the comparison of greedy heuristics for tree and rule construction as single-criterion and bi-criteria optimization algorithms; and the development of a restricted multi-pruning approach used in classification and knowledge representation.
Decision tree is a widely used form of representing algorithms and knowledge. Compact data models and fast algorithms require optimization of tree complexity. This book is a research monograph on average time complexity of decision trees. It generalizes several known results and considers a number of new problems. The book contains exact and approximate algorithms for decision tree optimization, and bounds on minimum average time complexity of decision trees. Methods of combinatorics, probability theory and complexity theory are used in the proofs as well as concepts from various branches of discrete mathematics and computer science. The considered applications include the study of average depth of decision trees for Boolean functions from closed classes, the comparison of results of the performance of greedy heuristics for average depth minimization with optimal decision trees constructed by dynamic programming algorithm, and optimization of decision trees for the corner point recognition problem from computer vision. The book can be interesting for researchers working on time complexity of algorithms and specialists in test theory, rough set theory, logical analysis of data and machine learning.
Dynamic programming is an efficient technique for solving optimization problems. It is based on breaking the initial problem down into simpler ones and solving these sub-problems, beginning with the simplest ones. A conventional dynamic programming algorithm returns an optimal object from a given set of objects. This book develops extensions of dynamic programming, enabling us to (i) describe the set of objects under consideration; (ii) perform a multi-stage optimization of objects relative to different criteria; (iii) count the number of optimal objects; (iv) find the set of Pareto optimal points for bi-criteria optimization problems; and (v) to study relationships between two criteria. It considers various applications, including optimization of decision trees and decision rule systems as algorithms for problem solving, as ways for knowledge representation, and as classifiers; optimization of element partition trees for rectangular meshes, which are used in finite element methods for solving PDEs; and multi-stage optimization for such classic combinatorial optimization problems as matrix chain multiplication, binary search trees, global sequence alignment, and shortest paths. The results presented are useful for researchers in combinatorial optimization, data mining, knowledge discovery, machine learning, and finite element methods, especially those working in rough set theory, test theory, logical analysis of data, and PDE solvers. This book can be used as the basis for graduate courses.
Decision tree is a widely used form of representing algorithms and knowledge. Compact data models and fast algorithms require optimization of tree complexity. This book is a research monograph on average time complexity of decision trees. It generalizes several known results and considers a number of new problems. The book contains exact and approximate algorithms for decision tree optimization, and bounds on minimum average time complexity of decision trees. Methods of combinatorics, probability theory and complexity theory are used in the proofs as well as concepts from various branches of discrete mathematics and computer science. The considered applications include the study of average depth of decision trees for Boolean functions from closed classes, the comparison of results of the performance of greedy heuristics for average depth minimization with optimal decision trees constructed by dynamic programming algorithm, and optimization of decision trees for the corner point recognition problem from computer vision. The book can be interesting for researchers working on time complexity of algorithms and specialists in test theory, rough set theory, logical analysis of data and machine learning.
In this book, the concept of a hypothesis about the values of all attributes is added to the standard decision tree model, considered, in particular, in test theory and rough set theory. This extension allows us to use the analog of equivalence queries from exact learning and explore decision trees that are based on various combinations of attributes, hypotheses, and proper hypotheses (analog of proper equivalence queries). The two main goals of this book are (i) to provide tools for the experimental and theoretical study of decision trees with hypotheses and (ii) to compare these decision trees with conventional decision trees that use only queries, each based on a single attribute. Both experimental and theoretical results show that decision trees with hypotheses can have less complexity than conventional decision trees. These results open up some prospects for using decision trees with hypotheses as a means of knowledge representation and algorithms for computing Boolean functions. The obtained theoretical results and tools for studying decision trees with hypotheses are useful for researchers using decision trees and rules in data analysis. This book can also be used as the basis for graduate courses.
Dynamic programming is an efficient technique for solving optimization problems. It is based on breaking the initial problem down into simpler ones and solving these sub-problems, beginning with the simplest ones. A conventional dynamic programming algorithm returns an optimal object from a given set of objects. This book develops extensions of dynamic programming, enabling us to (i) describe the set of objects under consideration; (ii) perform a multi-stage optimization of objects relative to different criteria; (iii) count the number of optimal objects; (iv) find the set of Pareto optimal points for bi-criteria optimization problems; and (v) to study relationships between two criteria. It considers various applications, including optimization of decision trees and decision rule systems as algorithms for problem solving, as ways for knowledge representation, and as classifiers; optimization of element partition trees for rectangular meshes, which are used in finite element methods for solving PDEs; and multi-stage optimization for such classic combinatorial optimization problems as matrix chain multiplication, binary search trees, global sequence alignment, and shortest paths. The results presented are useful for researchers in combinatorial optimization, data mining, knowledge discovery, machine learning, and finite element methods, especially those working in rough set theory, test theory, logical analysis of data, and PDE solvers. This book can be used as the basis for graduate courses.
The results presented here (including the assessment of a new tool – inhibitory trees) offer valuable tools for researchers in the areas of data mining, knowledge discovery, and machine learning, especially those whose work involves decision tables with many-valued decisions. The authors consider various examples of problems and corresponding decision tables with many-valued decisions, discuss the difference between decision and inhibitory trees and rules, and develop tools for their analysis and design. Applications include the study of totally optimal (optimal in relation to a number of criteria simultaneously) decision and inhibitory trees and rules; the comparison of greedy heuristics for tree and rule construction as single-criterion and bi-criteria optimization algorithms; and the development of a restricted multi-pruning approach used in classification and knowledge representation.
In this book, the following three approaches to data analysis are presented: - Test Theory, founded by Sergei V. Yablonskii (1924-1998); the first publications appeared in 1955 and 1958, - Rough Sets, founded by Zdzisław I. Pawlak (1926-2006); the first publications appeared in 1981 and 1982, - Logical Analysis of Data, founded by Peter L. Hammer (1936-2006); the first publications appeared in 1986 and 1988. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected. - Logical Analysis of Data, founded by Peter L. Hammer (1936-2006); the first publications appeared in 1986 and 1988. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected. These three approaches have much in common, but researchers active in one of these areas often have a limited knowledge about the results and methods developed in the other two. On the other hand, each of the approaches shows some originality and we believe that the exchange of knowledge can stimulate further development of each of them. This can lead to new theoretical results and real-life applications and, in particular, new results based on combination of these three data analysis approaches can be expected.
This will help us customize your experience to showcase the most relevant content to your age group
Please select from below
Login
Not registered?
Sign up
Already registered?
Success – Your message will goes here
We'd love to hear from you!
Thank you for visiting our website. Would you like to provide feedback on how we could improve your experience?
This site does not use any third party cookies with one exception — it uses cookies from Google to deliver its services and to analyze traffic.Learn More.