Piet MONDRIAN Brodway Boogie Woogie
Raymond.Balestra@ac-nice.fr. Piet MONDRIAN. Broadway Boogie-Woogie. Piet MONDRIAN / Broadway Boogie Woogie / 1942 – 1943 / Huile sur toile (127 x 127cm)
Poliform
8 nov. 2021 Wooden coffee tables and wooden backrests with shelf. Panels in wood and by-products veneered. MONDRIAN. Jean-Marie Massaud
Piet MONDRIAN est né le 7 mars 1872 en Hollande. En 1892 il a 20
Piet MONDRIAN est né le 7 mars 1872 en Hollande. En 1892 il a 20 ans et part étudier la peinture à Amsterdam. Il y fait la rencontre du poète français Paul
TOP SECRET Réservés aux élèves de la classe de 6 3
12 fév. 2014 Vous avez reproduit le tableau de Piet Mondrian sur une feuille blanche vous l'avez analysé dans tous les sens et sous toutes ses coutures.
The Mondrian Process in Machine Learning arXiv:1507.05181v1
18 juil. 2015 The Mondrian process has been used as the main building block of a clever online random forest classification algorithm that turns out to be ...
COMPANY OVERVIEW
Mondrian Doha debuts with the most decadent bridal suite in the world. MONDRIAN LA. Mondrian is the most happening place in LA. The Hotel Rooftop: Expensive
Mondrian Forests
Mondrian Forests. Randomization mechanism. Online training. Prediction and Hierarchical smoothing. Classification Experiments: online vs batch.
The Mondrian Process
Mondrian processes are random partitions on product spaces not constrained to be regular grids. Much like kd-trees. Mondrian processes partition a space
Le sens du désordre - Rencontre entre Vera Molnar et Piet Mondrian
Piet Mondrian. Vera Molnar. Amalgamer avec Scratch. PIET MONDRIAN. Artiste peintre né à la fin du XIXe siècle aux Pays-Bas. C'est une figure majeure de.
Mondrian Forests
Balaji Lakshminarayanan
Gatsby Unit, University College London
Joint work with Daniel M. Roy and Yee Whye Teh
1Outline
Motivation and Background
Mondrian Forests
Randomization mechanism
Online training
Prediction and Hierarchical smoothing
Classification Experiments: online vs batch
Regression Experiments: evaluating uncertainty estimatesConclusion
2Motivation
Typical converation:
I have a fasterABC DEF sampler for a fancy
non-parametric Bayesian model XYZBayesian: cool!
Others: Isn"t the
non-Ba yesianpar ametric v ersion,lik e100 times faster? Why should I care?Lots of neat ideas in Bayesian non-parametrics; can we use these in a non-Bayesian context?3Motivation
Typical converation:
I have a fasterABC DEF sampler for a fancy
non-parametric Bayesian model XYZBayesian: cool!
Others: Isn"t the
non-Ba yesianpar ametric v ersion,lik e100 times faster? Why should I care?Lots of neat ideas in Bayesian non-parametrics; can we use these in a non-Bayesian context?3Motivation
Typical converation:
I have a fasterABC DEF sampler for a fancy
non-parametric Bayesian model XYZBayesian: cool!
Others: Isn"t the
non-Ba yesianpar ametric v ersion,lik e100 times faster? Why should I care?Lots of neat ideas in Bayesian non-parametrics; can we use these in a non-Bayesian context?3Motivation
Typical converation:
I have a fasterABC DEF sampler for a fancy
non-parametric Bayesian model XYZBayesian: cool!
Others: Isn"t the
non-Ba yesianpar ametric v ersion,lik e100 times faster? Why should I care?Lots of neat ideas in Bayesian non-parametrics; can we use these in a non-Bayesian context?3Problem setup
Input: attributesX=fxngNn=1, labelsY=fyngNn=1(i.i.d) xn2 X(we assumeX= [0;1]Dbut could be more general) yn2 f1;:::;Kg(classification) oryn2R(regression)Goal: Predictyfor test datax
Recipe for prediction: Use ar andomf orest
Ensemb leof r andomizeddecision trees
Stat e-of-the-artf orlots of real w orldprediction tasks -Do we Need Hundreds of Classifiers to Solve Real World Classification Problems?[Fern´andez-Delgado et al., 2014]What is a decision tree?4
Problem setup
Input: attributesX=fxngNn=1, labelsY=fyngNn=1(i.i.d) xn2 X(we assumeX= [0;1]Dbut could be more general) yn2 f1;:::;Kg(classification) oryn2R(regression)Goal: Predictyfor test datax
Recipe for prediction: Use ar andomf orest
Ensemb leof r andomizeddecision trees
Stat e-of-the-artf orlots of real w orldprediction tasks -Do we Need Hundreds of Classifiers to Solve Real World Classification Problems?[Fern´andez-Delgado et al., 2014]What is a decision tree?4
Problem setup
Input: attributesX=fxngNn=1, labelsY=fyngNn=1(i.i.d) xn2 X(we assumeX= [0;1]Dbut could be more general) yn2 f1;:::;Kg(classification) oryn2R(regression)Goal: Predictyfor test datax
Recipe for prediction: Use ar andomf orest
Ensemb leof r andomizeddecision trees
Stat e-of-the-artf orlots of real w orldprediction tasks -Do we Need Hundreds of Classifiers to Solve Real World Classification Problems?[Fern´andez-Delgado et al., 2014]What is a decision tree?4
Example: Classification tree
Hierarchical axis-aligned binary partitioning of input spaceRule for predicting label within each blockx
1>0:37x
2>0:5 , ,F,F?
??x 2x101 1B jT: list of nodes, feature-id + location of splits for internal nodes : Multinomial parameters at leaf nodes 5Prediction using decision tree
Example
Multi- classclassification: = [r;b;g]
Predict ion= smoothed empir icalhistog ramin node jLabel counts in left node [nr=2;nb=0;ng=0]
- Dirichlet(=3;=3;=3)Predict ion= P osteriormean of =2+=32+;=32+;=32+?
??x 2x101 1B j6Prediction using decision tree
Example
Multi- classclassification: = [r;b;g]
Predict ion= smoothed empir icalhistog ramin node jLabel counts in left node [nr=2;nb=0;ng=0]
- Dirichlet(=3;=3;=3)Predict ion= P osteriormean of =2+=32+;=32+;=32+
Likelihood fornthdata point=p(ynjj)assumingxnlies in leaf nodejofTPrior overj: independent orhier archical
Prediction forxfalling inj=EjjT;X;Yp(yjj), where
p(jjT;X;Y)/p(jj:::)|{z} priorY n2N(j)p(ynjj) |{z} likelihood of data points in node jSmoothing is done independently for each tree6
From decision trees to Random forests (RF)
Generate
r andomized trees fTmgM1Prediction forx:
p(yjx) =1M X mp(yjx;Tm)Model combination
and not Ba yesianmodel a veragingAdvantages of RF
Excellent predictiv eperf ormance(test accur acy)
F astto tr ain(in batch setting) and test
T reescan be tr ainedin par allel7
From decision trees to Random forests (RF)
Generate
r andomized trees fTmgM1Prediction forx:
p(yjx) =1M X mp(yjx;Tm)Model combination
and not Ba yesianmodel a veragingAdvantages of RF
Excellent predictiv eperf ormance(test accur acy)
F astto tr ain(in batch setting) and test
T reescan be tr ainedin par allel7
Disadvantages of RF
Not possible to train incrementally
Re-tr ainingbatch v ersionper iodicallyis slo wO(N2logN)Existin gonline RF v ariants
Saffari et al., 2009
Denil et al., 2013
] require lots of memor y/ computation or need lots of tr ainingdata bef orethe ycan deliv ergood test accuracy ( data inefficient Random forests do not give useful uncertainty estimatesPredict ionsoutside r angeof tr ainingdata can be
overconfident Uncer taintyestimates are cr ucialin applications such as Bayesian optimization, Just-in-time learning, reinforcement learning, etc.8Disadvantages of RF
Not possible to train incrementally
Re-tr ainingbatch v ersionper iodicallyis slo wO(N2logN)Existin gonline RF v ariants
Saffari et al., 2009
Denil et al., 2013
] require lots of memor y/ computation or need lots of tr ainingdata bef orethe ycan deliv ergood test accuracy ( data inefficient Random forests do not give useful uncertainty estimatesPredict ionsoutside r angeof tr ainingdata can be
overconfident Uncer taintyestimates are cr ucialin applications such as Bayesian optimization, Just-in-time learning, reinforcement learning, etc.8Mondrian Forests
Mondrian forests= Mondrian process + Random forestsCan operate in either batch mode or online mode
Online speedO(NlogN)
Data efficient
(predictiv eperf ormanceof online mode equals that of batch mode!)Better uncertainty estimate than random forests
Predictions outside range of training data exhibit higher uncertainty and shrink to prior as you move farther away9Mondrian Forests
Mondrian forests= Mondrian process + Random forestsCan operate in either batch mode or online mode
Online speedO(NlogN)
Data efficient
(predictiv eperf ormanceof online mode equals that of batch mode!)Better uncertainty estimate than random forests
Predictions outside range of training data exhibit higher uncertainty and shrink to prior as you move farther away9Outline
Motivation and Background
Mondrian Forests
Randomization mechanism
Online training
Prediction and Hierarchical smoothing
Classification Experiments: online vs batch
Regression Experiments: evaluating uncertainty estimatesConclusion
10Popular batch RF variants
How to generate individual trees in RF?
Breiman-RF[Breiman, 2001]: Bagging + Randomly
subsample features and choose best location amongst subsampled featuresExtremely Randomized Trees[Geurts et al., 2006]
(ERT-k): Randomly samplek(feature-id, location) pairs and choose the best split amongst this subset no b aggingER T-1does not use labels Yto guide splits!11
Popular batch RF variants
How to generate individual trees in RF?
Breiman-RF[Breiman, 2001]: Bagging + Randomly
subsample features and choose best location amongst subsampled featuresExtremely Randomized Trees[Geurts et al., 2006]
(ERT-k): Randomly samplek(feature-id, location) pairs and choose the best split amongst this subset no b aggingER T-1does not use labels Yto guide splits!11
Mondrian process [Roy and Teh, 2009]
MP(;X)specifies a distribution over hierarchical
axis-aligned binary partitions ofX(e.g.RD,[0;1]D)is complexity parameter of the Mondrian processFigure:Mondrian Composition II in Red, Blue and Yellow (Source: Wikipedia)12
Mondrian process [Roy and Teh, 2009]
MP(;X)specifies a distribution over hierarchical
axis-aligned binary partitions ofX(e.g.RD,[0;1]D)is complexity parameter of the Mondrian processFigure:Mondrian Composition II in Red, Blue and Yellow (Source: Wikipedia)12
Generative process: MP(;f[`1;u1];[`2;u2]g)
1.Dr awfrom exponential with rateu1`1+u2`2
2.IF> stop,?
1 u 1 u 2 213Generative process: MP(;f[`1;u1];[`2;u2]g)
1.Dr awfrom exponential with rateu1`1+u2`2
2.IF> stop,ELSE, sample a split
sp litdimension: choose dimension dwith prob/ud`d sp litlocation: choose unif ormlyfrom [`d;ud]? 1 u 1 u 2 214Generative process: MP(;f[`1;u1];[`2;u2]g)
1.Dr awfrom exponential with rateu1`1+u2`2
2.IF> stop,ELSE, sample cut
Ch oosedimension dwith probability/ud`d
Ch oosecut location unif ormlyfrom [`d;ud]
Recu rseon left and r ightsubtrees with par ameter? 1 u 1 u 2 215Self-consistency of Mondrian process
SimulateT MP(;[`1;u1];[`2;u2])?
1 u 1 u 2 2 Restriction has distribution MP(;[`01;u01];[`02;u02])!16Self-consistency of Mondrian process
SimulateT MP(;[`1;u1];[`2;u2])
RestrictTto a smaller rectangle[`01;u01][`02;u02]? 1 u 1 u 2 2 Restriction has distribution MP(;[`01;u01];[`02;u02])!16Self-consistency of Mondrian process
SimulateT MP(;[`1;u1];[`2;u2])
RestrictTto a smaller rectangle[`01;u01][`02;u02]? 1 u 1 u 2 2 Restriction has distribution MP(;[`01;u01];[`02;u02])!16Mondrian trees
UseXto define lower and upper limits within each node and use MP to sample splits. Difference between Mondrian tree and usual decision tree split in node jis committed only within extent of training data in nodej node jis associated with 'time of split"tj>0 (split time increases with depth and will be useful in online training) splits are chosen independent of the labels Y -is 'weighted max-depth".x1>0:37x
2>0:5 , ,F,F-
-00:420:7∞?
??x 2x101 1B xj17Mondrian trees
UseXto define lower and upper limits within each node and use MP to sample splits. Difference between Mondrian tree and usual decision tree split in node jis committed only within extent of training data in nodej node jis associated with 'time of split"tj>0 (split time increases with depth and will be useful in online training) splits are chosen independent of the labels Y -is 'weighted max-depth".x1>0:37x
2>0:5 , ,F,F-
-00:420:7∞?
??x 2x101 1B xj17Outline
Motivation and Background
Mondrian Forests
Randomization mechanism
Online training
Prediction and Hierarchical smoothing
Classification Experiments: online vs batch
Regression Experiments: evaluating uncertainty estimatesConclusion
18Mondrian trees: online learning
As dataset grows, we extend the Mondrian treeTby
simulating from a conditional Mondr ianprocess MTxT MT(;D1:n) T0j T;D1:n+1MTx(;T;Dn+1)=) T0MT(;D1:n+1)
Distribution of batch and online trees are the same!Order of the data points does not matter
quotesdbs_dbs47.pdfusesText_47[PDF] monecole evaluation
[PDF] monet impression soleil levant cycle 3
[PDF] Money can't buy hapiness
[PDF] moneyness de l option
[PDF] monica's boots
[PDF] monks oliver twist
[PDF] Monnaie et métaux: devoir de Physique
[PDF] monnaie romaine synonyme
[PDF] monnet (le parlement de Londre)
[PDF] Monochrome
[PDF] monodie polyphonie college
[PDF] monodie sur bourdon definition
[PDF] monoeuvre avis
[PDF] monoeuvre code promo