Escolar Documentos
Profissional Documentos
Cultura Documentos
EBEB 2016
XIII Brazilian Meeting on Bayesian Statistics
February 2225, 2016 Belo Horizonte Brazil
Introduction I
In randomized experiments, a simple random allocation can
it has an exponential computational cost in the number of covariates (for fixed balance constraints).
Introduction II
We propose the use of Haphazard Intentional Allocation, an al-
ternative allocation method based on optimal balance of the covariates extended by random noise, see Lauretto et al. (2012).
Similarly to the allocation process in Morgan and Rubin (2012),
From now on, we consider the case of two groups, G = {0, 1},
function is based on the Mahalanobis distance between the covariates of interest in each group.
In order to define this loss function, let A be an arbitrary matrix
M (w, A) =
p
n1 n0 /n ka1 a0 k2
(1)
(2)
dka1 a0 k .
of Linear Programming:
minimize(w) H(, w, X, Z)
= H(w, Z) + (1 )H(w, X)
such that
w t 1 = n1
w {0, 1}n
(3)
Numerical Experiments I
In order to perform a haphazard intentional allocation, it is
Numerical Experiments II
Each method ran under a budget of 5, 10, 20, 60 and 300
(on the original data, that is, M (w, X) ) for the resulting allocations yielded by:
The Haphazard Intentional Allocation method optimizing the
XXX budget
XXX
procedure
XX
Haphazard, = 0.1
Haphazard, = 0.36
Haphazard, = 0.55
Haphazard, = 0.68
Haphazard, = 0.77
Rerandomization
Pure randomization
5s
10s
20s
60s
300s
.038
.045
.051
.057
.061
.217
.037
.044
.050
.054
.059
.209
.035
.041
.045
.049
.053
.203
0.462
.032
.039
.042
.045
.048
.192
.028
.033
.036
.038
.041
.178
Numerical Experiments IV
Table 1 suggests the following conclusions:
The larger the time budget, the smaller the median value of the
loss function M (w, X).
The smaller the value of , less noise is added to the optimization problem and, therefore, the smaller the median value of the
loss function M (w, X).
Choosing = 0.1, Haphazard Intentional Allocation obtains a
median Mahalanobis loss that is at least 5 times smaller than
when using the Rerandomization method.
Figures 1a, 1b illustrate the difference in covariate balance be-
Numerical Experiments V
vocabpre
mathpre
numbmath
likemath
likelit
preflit
actcomp
collgpaa
age
male
pconsc
pintell
HH
RR
Pure
Numerical Experiments VI
majormi
hsgpaar
married
momdegr
daddegr
credit
parentsIncome
beck
mars
pextra
pemot
pagree
HH
RR
Pure
Yule coefficient.
Haphazard, = 0.1
Haphazard, = 0.36
Haphazard, = 0.55
Haphazard, = 0.68
Haphazard, = 0.77
Rerandomization
Pure randomization
5s
10s
20s
60s
300s
.18
.17
.21
.22
.26
.10
.15
.14
.16
.17
.20
.10
.13
.12
.13
.12
.12
.10
.10
.12
.11
.11
.11
.11
.10
.11
.11
.11
.11
.11
.10
Numerical Experiments IX
Empirically, Rerandomization attains a Yule coefficient compa-
Numerical Experiments X
Numerical Experiments XI
method are uniformly more powerful over than the ones obtained using the Rerandomization method.
Figure 3 shows that the difference in power between these al-
1.0
Rerandomization
Haphazard
0.8
0.6
power
0.4
0.2
0.0
0.2
0.4
0.6
0.8
1.0
0.7
0.6
0.4
0.3
0.2
0.1
0.0
Difference in power
0.5
0.0
0.2
0.4
0.6
0.8
1.0
Figure 3. Difference between power curves of Haphazard and Rerandomization Allocations for testing = 0 using a permutation test.
Future Research
References I
[1] G.H. Golub and C.F. Van Loan. Matrix Computations. JHU Press, 2012.
[2] J. Y. Gotoh and S. Uryasev. Two pairs of polyhedral norms versus lp -norms:
proximity and applications in optimization. Mathematical Programming A,
156, 391431, 2016.
[3] Gurobi Optimization Inc. gurobi: Gurobi Optimizer 6.5 interface. URL
http://www.gurobi.com, 2015.
[4] M. S. Lauretto, F. Nakano, C. A. B. Pereira, J. M. Stern Intentional Sampling by Goal Optimization with Decoupling by Stochastic Perturbation. AIP
Conference Proceedings, 1490, 189-201, 2012.
[5] R. F. Love, J. G. Morris and G. O. Wesolowsky. Facilities location. Chapter
3:5160, 1988.
[6] R. K. Martin. Large scale linear and integer optimization: a unified approach.
Springer Science & Business Media, 2012.
[7] K. L. Morgan and D. B. Rubin. Rerandomization to improve covariate balance in experiments. The Annals of Statistics 40 (2), 12631282, 2012.
References II
[8] K. L. Morgan and D. B. Rubin. Rerandomization to Balance Tiers of Covariates. JASA, 110, 512, 14121421, 2015.
[9] B. A. Murtag. Advanced linear programming: computation and practice.
McGraw-Hill International Book, 1981
[10] K. Pavlikov and S. Uryasev. CVaR norm and applications in optimization.
Optimization Letters 8, 19992020, 2014.
[11] W. R. Shadish, M. R. Clark, P. M. Steiner. Can nonrandomized experiments
yield accurate answers? A randomized experiment comparing random and
nonrandom assignments. Journal of the American Statistical Association
103(484), 13341344, 2008.
[12] J. E. Ward and R. E. Wendell. Technical Note-A New Norm for Measuring
Distance Which Yields Linear Location Problems. Operations Research 28
(3-part-ii), 836844, 1980.
[13] J. E. Ward and R. E. Wendell. Using Block Norms for Location Modeling.
Operations Research, 33(5), 10741090, 1985.