Title: Results of the Causality Challenge
1Results of the Causality Challenge
- Isabelle Guyon, Clopinet
- Constantin Aliferis and Alexander Statnikov,
Vanderbilt Univ. - André Elisseeff and Jean-Philippe Pellet, IBM
Zürich - Gregory F. Cooper, Pittsburg University
- Peter Spirtes, Carnegie Mellon
2Causal discovery
What affects
- Which actions will have beneficial effects?
3Systemic causality
4Feature Selection
Y
X
Predict Y from features X1, X2, Select most
predictive features.
5Causation
Predict the consequences of actions Under
manipulations by an external agent, some
features are no longer predictive.
6Challenge Design
7Available data
- A lot of observational data.
- Correlation ? Causality!
- Experiments are often needed, but
- Costly
- Unethical
- Infeasible
- This challenge, semi-artificial data
- Re-simulated data
- Real data with artificial probes
8Four tasks
9On-line feed-back
10Difficulties
- Violated assumptions
- Causal sufficiency
- Markov equivalence
- Faithfulness
- Linearity
- Gaussianity
- Overfitting (statistical complexity)
- Finite sample size
- Algorithm efficiency (computational complexity)
- Thousands of variables
- Tens of thousands of examples
11Evaluation
- Fulfillment of an objective
- Prediction of a target variable
- Predictions under manipulations
- Causal relationships
- Existence
- Strength
- Degree
12Setting
- Predict a target variable (on training and test
data). - Return the set of features used.
- Flexibility
- Sorted or unsorted list of features
- Single prediction or table of results
- Complete entry xxx0, xxx1, xxx2 results (for at
least one dataset).
13Metrics
- Results ranked according to the test set
- target prediction performance Tscore
- We also assess directly the feature set with a
Fscore, not used for ranking.
14Toy Examples
15Causality assessmentwith manipulations
16LUCAS1 manipulated
Causality assessmentwith manipulations
17Causality assessmentwith manipulations
LUCAS2 manipulated
18Goal driven causality
- We define
- Vvariables of interest
- (e.g. MB, direct causes, ...)
- We assess causal relevance Fscoref(V,S).
19Causality assessmentwithout manipulation?
20Using artificial probes
Anxiety
Peer Pressure
Born an Even Day
Smoking
Genetics
Yellow Fingers
Lung Cancer
Attention Disorder
Allergy
LUCAP0 natural
Coughing
Fatigue
Car Accident
21Using artificial probes
LUCAP12 manipulated
22Scoring using probes
- What we can compute (Fscore)
- Negative class probes (here, all non-causes,
all manipulated). - Positive class other variables (may include
causes and non causes). - What we want (Rscore)
- Positive class causes.
- Negative class non-causes.
- What we get (asymptotically)
- Fscore (NTruePos/NReal) Rscore 0.5
(NTrueNeg/NReal)
23Results
24Challenge statistics
- Start December 15, 2007.
- End April 30, 2000
- Total duration 20 weeks.
- Last (complete) entry ranked
Number of ranked entrants
Number of ranked submissions
25Learning curves
26AUC distribution
27REGED
28SIDO
29CINA
30MARTI
31Pairwise comparisons
32Top ranking methods
- According to the rules of the challenge
- Yin Wen Chang SVM gt best prediction accuracy on
REGED and CINA. Prize 400 donated by Microsoft. - Gavin Cawley Causal explorer linear ridge
regression ensembles gt best prediction accuracy
on SIDO and MARTI. Prize 400 donated by
Microsoft. - According to pairwise comparisons
- Jianxin Yin and Prof. Zhi Gengs group Partial
Orientation and Local Structural Learning gt best
on Pareto front, new original causal discovery
algorithm. Prize free WCCI 2008 registration.
33Pairwise comparisons
REGED
SIDO
CINA
MARTI
34Conclusion
- We have found good correlation between causation
and prediction under manipulations. - Several algorithms have demonstrated
effectiveness of discovering causal
relationships. - We still need to investigate what makes then fail
in some cases. - We need to capitalize on the power of classical
feature selection methods.