Multi-directional Ensembles of
Regression and Classification Trees
What is the relationship of
Machine Learning to other sciences?
Given:
Derive:
N.b.:
Some of the red entries have to be known in order to achieve this!
This encompasses many kinds of ML
Correspond to various kinds of f
This encompasses many kinds of ML
Correspond to various X, Y
This encompasses many kinds of ML
Correspond to
feedback available to the algorithm
while learning f
Machine Learning is just another kind of mathematical tool, that can be used to address questions in science and engineering
We start from data and rely on algorithms, rather than searching for a reasonable f ourselves.
Nowhere?
Anomaly detection is hard,
precisely because
f is not easily defined
Many approaches exist,
2 key ideas:
``Flexibility matters.
Any truly intelligent system must not only possess the ability to solve a task, but must also exhibit considerable flexibility with regard to the task itself.''
But, now:
This is what we call a versatile model
Can we lift an ensemble of predictive models to reveal the general structure of a given dataset?
Discovering general structure is typically the domain of probabilistic methods,
i.e. Bayesian Networks.
Can it be done differently?
Can it be done by methods that are more suitable for big data?
+
-
Including every possible f is infeasible
1. We include a sample of possible trees in our ensemble
2. When presented with a prediction task,
we combine those trees appropriately
1. Make a selection of lego blocks
2. Build what you need with those lego blocks
There are 2 ways of building:
Select most appropriate trees, based on feature importance
Select most appropriate trees, based on feature importance
Select most appropriate trees, based on feature importance
Do Random Walks in the MERCS model
Do Random Walks in the MERCS model
Do Random Walks in the MERCS model
Do Random Walks in the MERCS model
Do Random Walks in the MERCS model
Machine Learning in spreadsheets
The task is to predict entries in a particular column from data present in the rest of the spreadsheet.
Missing data (i.e., empty cells) requires some degree of flexibility on the prediction algorithm to cope with changes in the available information (i.e., changing \(X\))
Anomaly detection for industrial applications.
Ideally, an algorithm designed to detect anomalies in this context needs to exhibit some degree of robustness against this kind of small, but common malfunctions.
Here, complex machines are monitored by many sensors and not every data source may function properly all the time.
Using the flexible predictions that MERCS allows to detect anomalies:
When we cannot predict something correctly,
this indicates an anomaly
A decision tree implicitly defines a hierarchical clustering
Use the structure as represented by the several clusters to detect anomalies