[PDF] [PDF] Applying Supervised Learning to Real-World Problems

The last years have seen machine learning methods applied to an increasing variety of application prob- lems such as: language, handwriting and speech pro-



Previous PDF Next PDF





[PDF] Applying Supervised Learning - MathWorks

A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains a model to generate 



Application of Supervised and Unsupervised Learning Methods to

application to fault diagnosis were studied: supervised and unsupervised learning methods Two types of neural networks based on supervised learning were 



[PDF] Supervised Learning Techniques in Mobile Device Apps for Androids

paradigms herein, supervised learning comprises situations where labeled training development in the area of machine learning for mobile apps Figure 2



[PDF] Applying Supervised Learning to Real-World Problems

The last years have seen machine learning methods applied to an increasing variety of application prob- lems such as: language, handwriting and speech pro-



[PDF] Applications of Supervised Machine Learning Algorithms in Additive

There are different supervised learning algorithms This work reviews and analyzes the applications of two supervised ML algorithms- support vector machine and 



[PDF] Supervised Machine Learning: A Review of - DataJobscom

The process of applying supervised ML to a real-world problem is described in Figure 1 Problem Data pre-processing Definition of training set Algorithm

[PDF] application of time value of money pdf

[PDF] application of vapour pressure

[PDF] application of word processing

[PDF] application of z transform in digital filters

[PDF] application of z transform in electrical engineering

[PDF] application of z transform in electronics and communication engineering

[PDF] application of z transform in engineering

[PDF] application of z transform in mechanical engineering

[PDF] application of z transform in real life

[PDF] application of z transform ppt

[PDF] application pour apprendre a compter

[PDF] application pour apprendre a compter ipad

[PDF] application pour apprendre a courir

[PDF] application pour apprendre a dessiner

[PDF] application pour apprendre a dessiner adulte

Applying Supervised Learning to Real-World Problems

Dragos D. Margineantu

Oregon State University

Department of Computer Science

303 Dearborn Hall

Corvallis, Oregon 97331-3202

margindr@cs.orst.edu

The last years have seen machine learning methods

applied to an increasing variety of application prob- lems such as: language, handwriting and speech pro- cessing, document classification, knowledge discovery in databases, industrial process control and diagnosis, fraud and intrusion detection, image analysis and many others.

Our work starts from the realization that most of

these problems require significant reformulation before learning algorithms can be applied, and in many cases, existing algorithms require modifications before being applied to a problem. The problems mentioned above differ in many aspects but, if subdivided into smaller problems (and this is the approach commonly taken), the sub-problems can often be formulated and approached by employing sim- ilar, unified supervised learning techniques. However, this divide-and-conquer process creates dependencies in the data that violate the assumption that the data are independent and identically distributed (iid) and that all errors are of equal cost, issues that involve making changes to existing learning algorithms. The purpose of my thesis is to identify some procedu- ral steps that are shared among the learning approaches to complex real-world problems, and to develop robust general purpose techniques and tools to replace ad-hoc decisions that are currently made during an application effort. The main topics my thesis will emphasize on are described as follows.

Learning with misclassification costs. Most

classification algorithms assume uniform class distri- bution and try to minimize the misclassification error. However, many applications require classifiers that min- imize an asymmetric loss function rather than the raw misclassification rate. In general, the cost of a wrong prediction depends both on the actual class and on the predicted class. One way to incorporate loss informa- tion into classifiers is to alter the priors based on labels of the training examples. For 2-class problems, this can be easily accomplished for any loss matrix, but for k > 2 classes, it is not sufficient. We have studied methods for setting the priors to best approximate an arbitrary k x k loss matrix in decision tree learners. We are currently studying alternative methods to incorporate loss infor- mation into decision trees, neural networks and other supervised learning algorithms.

Incorporating prior and common sense knowl-

edge and learning with non-independent data examples. Generally, besides the data available, there is simple, common sense knowledge about the data that is not made explicit in the dataset or database. One question is whether this knowledge is useful, and, if so, how it can be represented and communicated to a learning tool. Another question is what independence assumptions are made by the learning algorithm. The standard learning models and algorithms assume that the data is drawn independently and that it is identi- cally distributed. How is it possible to apply a learning method when this assumption is violated? One result that we have in this direction is that, if the data has a significant number of non-independent examples, en- semble learning methods like AdaBoost and Bagging do not improve performance (as expected) over single learners.

Determination of the granularity of the exam-

ples. As described in the previous section, the granu- larity of the examples is a common problem for super-vised learning approaches to different tasks. We believe that there are two distinct granularity problems to be solved for each general learning task: the granularity of the features (e.g. how large is the sliding window in a text-to-speech learning task?) and the granular- ity of the target labels (what do we want to predict, e.g. phonemes, stresses, or both?). Our work on this topic includes only some experiments on a manufactur- ing dataset.

Examples with multiple labels. The vast major-

ity of the research in supervised learning assumes that the labels of the examples are single values. However, there are cases when each example belongs to multiple classes. The training examples may be labeled either with single values or with a set of values. Possible ap- proaches to these kind of problems include: using learn- ing algorithms to output multiple values by exploiting the probability distributions, and using neural networks and to change the error function that is propagated back (as implemented in Apple's Newton character rec- ognizer).From: AAAI-99 Proceedings. Copyright © 1999, AAAI (www.aaai.org). All rights reserved.quotesdbs_dbs6.pdfusesText_12