[PDF] Multiscale Wavelets on Trees Graphs and High Dimensional Data





Previous PDF Next PDF



Machine Learning in Radiology: Applications Beyond Image

11 avr. 2017 Machine learning is a branch of artificial intelligence that has been employed in a variety of applications to analyze complex data sets and ...



Applying Supervised Learning

A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output) and trains a model to generate 



Application of Machine Learning to Beam Diagnostics

Abstract. Machine Learning (ML) techniques are widely used in science and industry to discover relevant information and make predictions from data.



Application of machine learning models and interpretability

To this end we use a flexible machine learning model



Runtime Verification for Critical Machine Learning Applications

In the last decade the application of Machine Learning (ML) has encountered an increasing interest in various applicative domains



Application of Machine Learning for the IPM-Based Profile

tool and machine learning algorithms lead to a beam profile correction methods for electron-collecting IPMs. INTRODUCTION. Ionization Profile Monitors (IPM) 



Multiscale Wavelets on Trees Graphs and High Dimensional Data

Theory and Applications to Semi Supervised Learning. Matan Gavish1 gavish@stanford.edu. Boaz Nadler boaz.nadler@weizmann.ac.il.



Application of Machine Learning To Epileptic Seizure Detection

We present and evaluate a machine learn- ing approach to constructing patient-specific classifiers that detect the onset of an epileptic seizure through 



Deep anomaly detection using self-supervised learning: application

10 mars 2022 detection using self-supervised learning: application to time series of cellular data. ASPAI 2021 -. 3rd International Conference on ...

Multiscale Wavelets on Trees, Graphs and High Dimensional Data: Theory and Applications to Semi Supervised Learning

Matan Gavish1gavish@stanford.edu

Boaz Nadlerboaz.nadler@weizmann.ac.il

Weizmann Institute of Science, P.O. Box 26, Rehovot, 76100, Israel

Ronald R. Coifmanronald.coifman@yale.edu

Yale University, New Haven, CT, 06520, USA

Abstract

Harmonic analysis, and in particular the re-

lation between function smoothness and ap- proximate sparsity of its wavelet coefficients, has played a key role in signal processing and statistical inference for low dimensional data. In contrast, harmonic analysis has thus far had little impact in modern problems in- volving high dimensional data, or data en- coded as graphs or networks. The main con- tribution of this paper is the development of a harmonic analysis approach, including both learning algorithms and supporting the- ory, applicable to these more general settings.

Given data (be it high dimensional, graph or

network) that is represented by one or more hierarchical trees, we first construct multi- scale wavelet-like orthonormal bases on it.

Second, we prove that in analogy to the Eu-

clidean case, function smoothness with re- spect to a specific metric induced by the tree is equivalent to exponential rate of coefficient decay, that is, to approximate sparsity. These results readily translate to simple practical algorithms for various learning tasks. We present an application to transductive semi- supervised learning.

1. Introduction

In recent years, vast data sets in the form of (i) graphs or networks, and (ii) data in high dimensional Eu-

1Currently at Stanford University, Stanford, CA.

Appearing inProceedings of the 27thInternational Confer- ence on Machine Learning, Haifa, Israel, 2010. Copyright

2010 by the author(s)/owner(s).

clidean space, are routinely collected in many areas. Analysis of these types of data is a major challenge for the statistics and machine learning communities. The statistical theory underlying data analysis has been studied extensively in both communities. The traditional statistics literature has, by and large, fo- cused on the setting of data in (low dimensional) Eu- clidean space (Lehmann & Casella, 2003; Hardle et al.,

1998). In contrast, the theoretical machine learning

community has developed a theory of learning from abstract hypothesis classes (Vapnik, 1998), where no- tions of function smoothness and the geometry of the ambient space are often absent. Neither of these traditional approaches is particu- larly well suited to the graph or high-dimensional Euclidean settings. Many traditional statistical in- ference approaches are inapplicable for high dimen- sional Euclidean data and become meaningless on non-

Euclidean data such as a graph. Nonetheless, both

graph data and high dimensional data typically have arich geometrical structure, which is often not directly exploited in the abstract machine learning framework. New statistical learning tools are thus needed, which would capitalize on the geometrical structure available in these settings. These tools should be accompanied by an underlying theory, specifying the conditions un- der which they can be expected to be useful. In this work we propose a harmonic analysis frame- work and develop such tools and supporting theory, under the assumption that the geometry of a graph or a high-dimensional Euclidean data set is captured by ahierarchical treeof increasingly refined partitions. We present two main contributions. First, given data encoded as a hierarchical tree, we build data adaptive wavelet-like orthonormal bases for the space of func- tions over the data set, in the spirit of the Haar basis on the unit interval [0,1] (see Fig. 1). Second, we Wavelets on trees, graphs and high dimensional data

Figure 1.An illustration of a Haar-like basis.

prove that for these bases, function smoothness with respect to a certain tree metric can be measured by the rate of function coefficient decay. In particular, for a balanced tree (defined below), smooth functions have coefficient decayexponentialin the tree level. As fast coefficient decay, or approximate sparsity, is the key principle underlying many inference techniques in the Euclidean case, our results readily translate into new simple and practical algorithms for various learn- ing tasks in the more general setting of graphs and high dimensional data. As a particular application of our approach, we present a novel semi-supervised learning (SSL) scheme. Preliminary results on various datasets show that our scheme is competitive with other ap- proaches, often achieving lower prediction errors.

2. Problem Setting and Related Work

LetX={x1,...,xN}be the dataset we wish to an-

alyze. The samplesximay be points in a high di- mensional space, or nodes in a weighted graph or net- work. Most supervised learning problems in this set-

ting involve inference on an unknown target functionfdefined onX. Examples include i)function denois-

ing: given valuesyi=f(xi) +?i, where?iare i.i.d. noise variables, estimate the underlyingf, ii)semi- supervised learning: givenf|SwhereSis a subset ofX, estimatef|X\S, iii)active learning: givenf|S choose the next unlabeled pointx?X\Sto query, to "optimally" estimatefwith as few queries as possible.

The key question is thus what are good methods

to represent, process and learn functions on general datasetsXhaving the form of a weighted graph or points in a high dimensional space.

In the ML community, a common approach to handle

high dimensional data is to represent it as a symmet- ric weighted graph with the aid of a similarity ker- nel. Common methods to process functions defined on graphs, in turn, are based on the graph Laplacian matrixL, and its variants (Chapelle et al., 2006). For example, in (Zhu et al., 2003) global function smooth- ness w.r.t. the graph is measured byfTLf. A related approach involves representing the target function in the eigenbasis of the graph Laplacian (Belkin & Niyogi,

2003). Here learning is performed by estimating the

first few eigenbasis coefficients from the labeled data. Despite their empirical success on various datasets, these global methods have several limitations. First, as described in (Nadler et al., 2009), measuring func- tion smoothness via the graph Laplacian (fTLf) leads to ill-posed problems for semi-supervised learning with high dimensional data as the number of unlabeled data grows to infinity. Second, eigenvector based methods are not always best suited for representing functions on a graph. Since these basis vectors are eigenvectors of a symmetric matrix, in general they haveglobalsupport and become increasinglyoscillatory. This limits the number of coefficients that can be robustly estimated. On the theoretical side, there is still no sufficient jus- tification for these methods for data that is not neces- sarily sampled from a low dimensional manifold. Fur- thermore, to date, even in the manifold setting there is no well understood theory for how many coefficients to estimate. For example, Belkin & Niyogi (2003), heuris- tically propose to estimaten/5 coefficients wherenis the total number of labeled points. Inspired by the classical Euclidean setting, where sim- ilar limitations of the Fourier basis are alleviated by wavelet bases, in this paper we propose amultiscale harmonic analysis approach to learning from graph or high dimensional data and develop both new algo- rithms as well as supporting theory. Our key assump- tion is that the geometry and structures of the input graph or high dimensional data are captured by one or several (possibly random) hierarchical trees. First, Wavelets on trees, graphs and high dimensional data we remark that trees are ubiquitous in statistics and computer science. Given a datasetX, there are many methods to construct a hierarchical tree, including de- terministic, random, agglomerative and divisive. Fur- thermore, in a Euclidean setting, tree-based classifiers are highly successful (Breiman et al., 1984; Breiman,

2001; Binev et al., 2005). Note, however, that our

setting is different as we do not necessarily assume a Euclidean structure. In this paper we do not focus on the exact method of tree construction, but rather assume that the input dataXis equipped with a hi- erarchical tree, either already given or constructed by some method. Our main results are a sensible defini- tion of function smoothness with respect to this tree, a corresponding multiscale learning approach and its supporting theory, assuming this function smoothness. As mentioned above, harmonic analysis has had thus far little impact on learning from graphs or from high dimensional data. As such, there are relatively few works suggesting multiscale representations for learn- ing in these settings (Coifman & Maggioni, 2006;

Mahadevan & Maggioni, 2006; Jansen et al., 2009).

Jansen et al. (2009) explicitly state the lack of support- ing theory for their methods: "The main disadvantage is that, apart from analogies with regular wavelets, there is currently no substantial body of theory be- hind our methods". This work provides a step forward towards the development of such a theory.

3. Main Results

LetXbe the given dataset, and letf:X→Rbe the

unknown target function to be learned. Further, de- note byV={f|f:X→R}the space of all functions on the dataset, with the inner product ?f,g?=1 NN j=1f(xj)g(xj).(1) Inspired by the success of multiscale wavelet decom- positions for 1-d signals and 2-d images (Mallat, 1999; Hardle et al., 1998), our harmonic analysis approach consists of constructing a multiscale basis{φj}N j=1for the spaceV, such that the target functionfadmits an efficient sparse representation in this basis. Accurate approximation offis then possible by estimating only a few of its coefficients?f,φj?.

3.1. Trees and Multi-Resulution Analysis

The starting point for constructing a basis forVis ahierarchical treerepresentation of the dataX. We denote by?= 1,...,Lthe level in the tree with?= 1

being the root and?=Lthe lowest level, where eachsamplexjis a single leaf. We further denote byX?kthe

set of all leaves of thek-th folder (subtree) of the tree at level?, and bysub(?,k) the number of subfolders of X ?kat the next level?+ 1. The crucial property we require of the given dataset Xis that the resulting tree isbalanced: that is, for all parent folders in the tree, 0< B

This property implies thatL=O(logN). Note that

B =B= 1/2 gives a perfectly balanced binary tree. Our next step is based on the following simple ob- servation (see also Donoho (1997); Murtagh (2007); Lee et al. (2008)):A tree representation of data nat- urally induces a multi-resolution analysis (MRA) with an associated Haar-like wavelet basis.In more detail, for each level?denote byV?the space of functions constant on all folders (subtrees) at level?, V ?={f|f:X→R,fconstant on all foldersX?j}. For example,V1is the one-dimensional space of con- stant functions on the dataset,V1= SpanR{1X}, whereasVL=V. By construction, V

1?...?V??V?+1?...?VL=V.(3)

This sequence of subspaces resembles a multiresolution analysis ofV, a key property for the development of wavelets on Euclidean spaces (Mallat, 1999). As in classical MRA, letW?(1?? < L) be the orthogonal complement ofV?inV?+1(V?+1=W??V?). The space of all functionsVcan then be decomposed as

V=VL=?L-1?

?=1W ???V1.(4)

3.2. Haar-like bases on the dataset

Eq. (4) is the key to construct multiscale, localized orthonormal bases for the spaceV. Consider a folder X ?kat level?that is split into two subfoldersX?+1 iand X ?+1 j. Then, there is a zero mean Haar-like function ?,k,1supported only on these two subfolders, which is piecewise constant on each of them (see for example ψ2,1,1in Fig. 1). If a folderX?kis split into three or more subfolders, thensub(?,k)-1 Haar-like orthonor- mal functions{ψ?,k,j}sub(?,k)-1 j=1need to be constructed. The collection of all these functions, augmented by the constant function onX, forms an orthonormal basis ofV, that we term aHaar-likebasis,B= {ψ?,k,j}. To clarify notation,?denotes the level of Wavelets on trees, graphs and high dimensional data the tree,kis the index of folderX?kat level?, and j= 1,...,sub(?,k)-1. For binary trees the third in- dex is not required, and we recover the standard dou- ble index wavelet notationψ?,k. Figure 1 illustrates some Haar-like wavelet functions for a simple tree. These functions resemble the classical Haar functions in the following sense: i) SinceW??V?+1, eachψ?,k,j is piecewise constant on folders at level?+1; ii) Since ?,k,jis supported on the folderX?k, it is nonzero only on folders at level?+ 1 which are subfolders ofX?k; iii) SinceW??V?, eachψ?,k,jis orthogonal to the constant function onX?k,?ψ?,k,j,1X?k?= 0.

3.3. Function Smoothness, Exponential

Coefficient Decay and Learnability

Our main theoretical result is an explicit relation be- tween the following concepts: thegeometryof the data, as represented by the tree structure; thesmoothness of a function with respect to the tree; and thedecay or sparsityof its coefficients in a Haar-like expansion. The practical implications of these theorems to func- tion learnability, e.g., to accurate approximation off from partially labeled data are considered in section 4. To relate the geometry of the data to the tree struc- ture, we first define a probability measureνon the tree as follows: For each setS?X, defineν(S) =|S|/|X|. Next, consider the following tree ultrametric, also pop- ular for metric approximations (Bartal, 1996; 1998):

Definition 3.1Tree Metric: For anyx,y?X, define

d(x,y) =?ν(folder(x,y))x?=y

0x=y(5)

wherefolder(x,y)is the smallest folder in the tree containing bothx,y. Given the tree metric, the following definition of smoothness is a straightforward analogue of H¨older smoothness in the Euclidean setting 1: Definition 3.2A functionfis(C,α)-H¨older w.r.t. the tree (with0< α <1) if With these definitions, the following theorems relate function smoothness, the geometry of the data and fast coefficient decay: Theorem 1Letf:X→Rbe(C,α)-H¨older and let ?,k,jbe a Haar-like basis function supported on the

1For trees constructed on Euclidean spaces, H¨older

w.r.t. the tree is not equivalent to H¨older in the Euclidean space. This issue is beyond the scope of this paper. folderX?k. Then |?f , ψ?,k,j?|?C2α+1·ν(X?k)α+1/2.(7)

The tree-balance requirement (2) implies that

ν(X?k)?

B?-1for all?,k. Therefore, smoothness

offimplies exponential decay rate of its wavelet coef- ficients as a function of the tree level?, |?f , ψ?,k,j?|?2α+1C·

B(?-1)(α+1/2).(8)

In particular, for a perfectly balanced binary tree with B= 1/2 we recover the familiar wavelet coefficient exponential rate of decay for smooth functions, see for example (Mallat, 1999), theorem 6.3.

Theorem 2Letf:X→R. Suppose that for all

functionsψ?,k,jin some Haar-like basis |?f,ψ?,k,j?|?C·ν(X?k)α+1/2 with someα >0, and some constantC. Thenfis (C?,α)-H¨older, withC?=2C

B3/211-Bα.

Finally, the following theorem, which seems new even in the Euclidean setting, shows an interesting relation betweenL1sparsity in a multiscale Haar-like basis, and function approximation:

Theorem 3LethI(x)be a Haar-like basis, where

each functionhIis supported on a setI?Xand

IaIhI(x).

Assume?

approximation

ˆf=?

|I|>?aIhI. Then ?f-ˆf?1=? ?.(9) Proofs of these theorems appear in the supplementary material. The key point of these theorems is that the multiscale representation of a weighted graph via a hi- erarchical tree allows for the development of a theory of harmonic analysis on graphs. Theorems 1-2 pro- vide a clear connection between the notion of function smoothness w.r.t. the tree and fast coefficient decay in a Haar-like expansion. These theorems have well known analogues in the Euclidean setting. Theorem 3 has interesting implications to statistical inference as it shows that under anL1bound on the coefficients in a Haar-like expansion, for a reconstruction error of

O(⎷

?) it is sufficient to estimate only coefficients cor- responding to Haar-like functions with support larger than?, that is, at mostO(1/?) coarse-scale coefficients.

Eq. (20) in supplementary material shows that our

Wavelets on trees, graphs and high dimensional data Haar-like functions satisfy the condition of Theorem

ν(X?k))1/2.

We emphasize that a balanced tree is a key assump- tion in our method. Empirically, on many datasets with given affinities, various graph-to-tree construc- tions indeed resulted in balanced trees. We note that the ability to represent data by a balanced tree is re- lated to some notion of "intrinsic low dimensionality" of the data. Theoretical conditions on graph affinities, on graph-to-tree constructions that provide balanced trees, and on the relations between function smooth- ness w.r.t. graph metrics and smoothness w.r.t. tree metrics are all subjects for further research. From a computational view, Haar-like functions are piecewise constant and thus simple to handle. Fur- thermore, expanding a function or estimating its coef- ficients in the Haar-like domain admit fast algorithms resembling the fast wavelet transform. One limitation of Haar-like functions is their lack of smoothness. Due to the arbitrary locations of their discontinuities, this may introduce artifacts when processing functions in the coefficient domain. In the context of signal denois- ing, (Coifman & Donoho, 1995) suggested to remove such artifacts by averaging shifted Haar bases. Simi- lar to their approach and to random forest (Breiman,

2001), we also average functions learned using several

quotesdbs_dbs6.pdfusesText_12
[PDF] application of time value of money pdf

[PDF] application of vapour pressure

[PDF] application of word processing

[PDF] application of z transform in digital filters

[PDF] application of z transform in electrical engineering

[PDF] application of z transform in engineering

[PDF] application of z transform in mechanical engineering

[PDF] application of z transform in real life

[PDF] application of z transform ppt

[PDF] application pour apprendre a compter

[PDF] application pour apprendre a compter ipad

[PDF] application pour apprendre a dessiner

[PDF] application pour apprendre a dessiner adulte

[PDF] application pour apprendre a dessiner ipad pro

[PDF] application pour apprendre a dessiner iphone