indirect multivariate response linear...

22
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 Biometrika (2016), xx, x, pp. 1–22 C 2007 Biometrika Trust Printed in Great Britain Indirect multivariate response linear regression BY AARON J. MOLSTAD AND ADAM J. ROTHMAN School of Statistics, University of Minnesota, Minneapolis, Minnesota 55455, U.S.A. [email protected] [email protected] SUMMARY We propose a class of estimators of the multivariate response linear regression coefficient matrix that exploits the assumption that the response and predictors have a joint multivariate nor- mal distribution. This allows us to indirectly estimate the regression coefficient matrix through shrinkage estimation of the parameters of the inverse regression, or the conditional distribution of the predictors given the responses. We establish a convergence rate bound for estimators in our class and we study two examples, which respectively assume that the inverse regression’s coefficient matrix is sparse and rank deficient. These estimators do not require that the forward regression coefficient matrix is sparse or has small Frobenius norm. Using simulation studies, we show that our estimators outperform competitors. Some key words: Covariance estimation; Reduced rank regression; Sparsity. 1. I NTRODUCTION Some statistical applications require the modeling of a multivariate response. Let y i R q be the measurement of the q-variate response for the ith subject and let x i R p be the nonrandom values of the p predictors for the ith subject (i =1,...,n). The multivariate response linear

Upload: others

Post on 09-Oct-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

Biometrika (2016), xx, x, pp. 1–22

C© 2007 Biometrika Trust

Printed in Great Britain

Indirect multivariate response linear regression

BY AARON J. MOLSTAD AND ADAM J. ROTHMAN

School of Statistics, University of Minnesota, Minneapolis, Minnesota 55455, U.S.A.

[email protected] [email protected]

SUMMARY

We propose a class of estimators of the multivariate response linear regression coefficient

matrix that exploits the assumption that the response and predictors have a joint multivariate nor-

mal distribution. This allows us to indirectly estimate the regression coefficient matrix through

shrinkage estimation of the parameters of the inverse regression, or the conditional distribution

of the predictors given the responses. We establish a convergence rate bound for estimators in

our class and we study two examples, which respectively assume that the inverse regression’s

coefficient matrix is sparse and rank deficient. These estimators do not require that the forward

regression coefficient matrix is sparse or has small Frobenius norm. Using simulation studies,

we show that our estimators outperform competitors.

Some key words: Covariance estimation; Reduced rank regression; Sparsity.

1. INTRODUCTION

Some statistical applications require the modeling of a multivariate response. Let yi ∈ Rq be

the measurement of the q-variate response for the ith subject and let xi ∈ Rp be the nonrandom

values of the p predictors for the ith subject (i = 1, . . . , n). The multivariate response linear

Page 2: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

81

82

83

84

85

86

87

88

89

90

91

92

93

94

95

96

2 AARON J. MOLSTAD AND ADAM J. ROTHMAN

regression model assumes that yi is a realization of the random vector

Yi = µ∗ + βT∗ xi + εi, i = 1, . . . , n, (1)

where µ∗ ∈ Rq is the unknown intercept, β∗ is the unknown p by q regression coefficient matrix,

and ε1, . . . , εn are independent copies of a mean zero random vector with covariance matrix

Σ∗E .

The ordinary least squares estimator of β∗ is

βOLS = arg minβ∈Rp×q

‖Y− Xβ‖2F , (2)

where ‖ · ‖F is the Frobenius norm, Rp×q is the set of real valued p by q matrices, Y is the

n by q matrix with ith row (Yi − n−1∑n

i=1 Yi)T, and X is the n by p matrix with ith row

(xi − n−1∑n

i=1 xi)T (i = 1, ..., n). It is well known that βOLS is the maximum likelihood es-

timator of β∗ when ε1, . . . , εn are independent and identically distributed Nq(0,Σ∗E) and the

corresponding maximum likelihood estimator of Σ−1∗E exists.

Many shrinkage estimators of β∗ have been proposed by penalizing the optimization in (2).

Some simultaneously estimate β∗ and remove irrelevant predictors (Turlach et al., 2005; Obozin-

ski et al., 2010; Peng et al., 2010). Others encourage an estimator of reduced rank (Yuan et al.,

2007; Chen & Huang, 2012).

Under the restriction that ε1, . . . , εn are independent and identically distributed Nq(0,Σ∗E),

shrinkage estimators of β∗ that penalize or constrain the minimization of the negative loglike-

lihood have been proposed. These methods simultaneously estimate β∗ and Σ−1∗E . Examples in-

clude maximum likelihood reduced-rank regression (Izenman, 1975; Reinsel & Velu, 1998),

envelope models (Cook et al., 2010; Su & Cook, 2011, 2012, 2013), and multivariate regression

with covariance estimation (Rothman et al., 2010; Lee & Liu, 2012; Bhadra & Mallick, 2013).

Page 3: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

Indirect multivariate response linear regression 3

To fit (1) with these shrinkage estimators, one exploits explicit assumptions about β∗ that

may be unreasonable in some applications. As an alternative, we propose an indirect method to

fit (1) without such assumptions. We instead assume that response and predictors have a joint

multivariate normal distribution and we employ shrinkage estimators of the parameters of the

conditional distribution of the predictors given the response. Our method provides alternative

indirect estimators of β∗, which may be suitable when existing shrinkage estimators are inad-

equate. In the very challenging case when p is large and β∗ is not sparse, one of our proposed

indirect estimators can predict well and be interpreted easily.

2. A NEW CLASS OF INDIRECT ESTIMATORS OF β∗

2·1. Class definition

We assume that the measured predictor and response pairs (x1, y1), . . . , (xn, yn) are a real-

ization of n independent copies of (X,Y ), where (XT, Y T)T ∼ Np+q(µ∗,Σ∗). We also assume

that Σ∗ is positive definite. Define the marginal parameters through the following partitions:

µ∗ =

µ∗Xµ∗Y

, Σ∗ =

Σ∗XX Σ∗XY

ΣT∗XY Σ∗Y Y

.

Our goal is to estimate the multivariate regression coefficient matrix β∗ = Σ−1∗XXΣ∗XY in the

forward regression model

Y | X = x ∼ Nq

µ∗Y + βT

∗ (x− µ∗X) ,Σ∗E,

without assuming that β∗ is sparse or that ‖β∗‖2F is small. To do this we will estimate the inverse

regression’s coefficient matrix η∗ = Σ−1∗Y Y ΣT

∗XY and the inverse regression’s error precision ma-

trix ∆−1∗ in the inverse regression model

X | Y = y ∼ Np

µ∗X + ηT

∗ (y − µ∗Y ) ,∆∗.

Page 4: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

166

167

168

169

170

171

172

173

174

175

176

177

178

179

180

181

182

183

184

185

186

187

188

189

190

191

192

4 AARON J. MOLSTAD AND ADAM J. ROTHMAN

We connect the parameters of the inverse regression model to β∗ with the following proposition,

which we prove in the Appendix.

PROPOSITION 1. If Σ∗ is positive definite, then

β∗ = ∆−1∗ ηT∗(Σ−1∗Y Y + η∗∆

−1∗ ηT∗)−1

. (3)

This result leads us to propose a class of estimators of β∗,

β = ∆−1ηT(Σ−1Y Y + η∆−1ηT)−1, (4)

where η, ∆, and ΣY Y are user-selected estimators of η∗, ∆∗, and Σ∗Y Y . If n > p+ q + 1 and

the ordinary sample estimators are used for η, ∆ and ΣY Y , then β is equivalent to βOLS.

We propose to use shrinkage estimators of η∗, ∆−1∗ , and Σ−1

∗Y Y in (4). This gives us the po-

tential to indirectly fit an unparsimonious forward regression model by fitting a parsimonious

inverse regression model. For example, η∗ could have only one nonzero entry and all entries in

β∗ could be nonzero. Conveniently, entries in η∗ can be interpreted like entries in β∗ are by re-

versing the roles of the predictors and responses. To fit the inverse regression model, we could

use any of the forward regression shrinkage estimators discussed in Section 1.

2·2. Related work

Lee & Liu (2012) proposed an estimator of β∗ that also exploits the assumption that

(XT, Y T)T is multivariate normal; however, unlike our approach which makes no explicit as-

sumptions about β∗, they assume that both Σ−1∗ and β∗ are sparse.

Modeling the inverse regression is a well-known idea in multivariate analysis. For example,

when Y is categorical, quadratic discriminant analysis treats X | Y = y as p-variate normal.

There are also many examples of modeling the inverse regression in the sufficient dimension

reduction literature (Adragni & Cook, 2009).

Page 5: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

193

194

195

196

197

198

199

200

201

202

203

204

205

206

207

208

209

210

211

212

213

214

215

216

217

218

219

220

221

222

223

224

225

226

227

228

229

230

231

232

233

234

235

236

237

238

239

240

Indirect multivariate response linear regression 5

The work most closely related to ours is Cook et al. (2013). They proposed indirect estimators

of β∗ based on modeling the inverse regression in the special case when the response is univariate,

i.e., q = 1. Under our multivariate normal assumption on (XT, Y T)T, Cook et al. (2013) showed

that

β∗ =1

1 + ΣT∗XY ∆−1

∗ Σ∗XY /Σ∗Y Y∆−1∗ Σ∗XY , (5)

and proposed estimators of β∗ by replacing Σ∗XY and Σ∗Y Y in (5) with their usual sample esti-

mators, and by replacing ∆−1∗ with a shrinkage estimator. This class of estimators was designed

to exploit an abundant signal rate in the forward univariate response regression when p > n.

3. ASYMPTOTIC ANALYSIS

We present a convergence rate bound for the indirect estimator of β∗ defined by (4). Our bound

allows p and q to grow with the sample size n. In the following proposition, ‖ · ‖ is the spectral

norm and ϕmin(·) is the minimum eigenvalue.

PROPOSITION 2. Suppose that the following conditions are true: (i) Σ∗ is positive definite for

all p+ q; (ii) the estimator Σ−1Y Y is positive definite for all q; (iii) the estimator ∆−1 is positive

definite for all p; (iv) there exists a positive constantK such that ϕmin(Σ−1∗Y Y ) ≥ K for all q; and

(v) there exist sequences an, bn and cn such that ‖η − η∗‖ = OP (an), ‖∆−1 −∆−1∗ ‖ =

OP (bn), ‖Σ−1Y Y − Σ−1

∗Y Y ‖ = OP (cn), and an‖η∗‖‖∆−1∗ ‖+ bn‖η∗‖2 + cn → 0 as n→∞. Then

‖β − β∗‖ = OP(an‖η∗‖2‖∆−1

∗ ‖2 + bn‖η∗‖3‖∆−1∗ ‖+ cn‖η∗‖‖∆−1

∗ ‖).

We prove Proposition 2 in the Supplementary Material. We used the spectral norm because it is

compatible with the convergence rate bounds established for sparse inverse covariance estimators

(Rothman et al., 2008; Lam & Fan, 2009; Ravikumar et al., 2011).

Page 6: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

241

242

243

244

245

246

247

248

249

250

251

252

253

254

255

256

257

258

259

260

261

262

263

264

265

266

267

268

269

270

271

272

273

274

275

276

277

278

279

280

281

282

283

284

285

286

287

288

6 AARON J. MOLSTAD AND ADAM J. ROTHMAN

If the inverse regression is parsimonious in the sense that ‖η∗‖ and ‖∆−1∗ ‖ are bounded,

then the bound in Proposition 2 simplifies to ‖β − β∗‖ = OP (an + bn + cn). We explore finite-

sample performance in Section 5.

4. ESTIMATORS IN OUR CLASS

4·1. Sparse inverse regression

We now describe an estimator of the forward regression coefficient matrix β∗ defined by (4)

that exploits zeros in the inverse regression’s coefficient matrix η∗, zeros in the inverse regres-

sion’s error precision matrix ∆−1∗ , and zeros in the precision matrix of the responses Σ−1

∗Y Y . We

estimate η∗ with

ηL1 = arg minη∈Rq×p

‖X− Yη‖2F +

p∑j=1

λj

q∑m=1

|ηmj |

, (6)

which separates into p L1-penalized least-squares regressions (Tibshirani, 1996): the first pre-

dictor regressed on the response through the pth predictor regressed on the response. We select

λj with 5-fold cross-validation, minimizing squared prediction error totaled over the folds, in

the regression of the jth predictor on the response (j = 1, . . . , p). This allows us to estimate the

columns of η∗ in parallel.

We estimate ∆−1∗ and Σ−1

∗Y Y with L1-penalized normal likelihood precision matrix estimation

(Yuan & Lin, 2007; Banerjee et al., 2008). Let Σ−1γ,S be a generic version of this estimator with

tuning parameter γ and input p by p sample covariance matrix S:

Σ−1γ,S = arg min

Ω∈Sp+

tr(ΩS)− log |Ω|+ γ∑j 6=k|ωjk|

, (7)

where Sp+ is the set of symmetric and positive definite p by p matrices. The optimization in (7)

was used to estimate the inverse regression’s error precision matrix in the univariate response

regression methods proposed by Cook et al. (2012) and Cook et al. (2013). There are many al-

Page 7: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

289

290

291

292

293

294

295

296

297

298

299

300

301

302

303

304

305

306

307

308

309

310

311

312

313

314

315

316

317

318

319

320

321

322

323

324

325

326

327

328

329

330

331

332

333

334

335

336

Indirect multivariate response linear regression 7

gorithms that solve (7). Two good choices are the graphical lasso (Yuan, 2008; Friedman et al.,

2008) and the algorithm of Hsieh et al. (2011). We select γ with 5-fold cross-validation maxi-

mizing a validation likelihood criterion (Huang et al., 2006):

γ = arg minγ∈G

5∑k=1

tr(

Σ−1γ,S(−k)

S(k)

)− log

∣∣∣Σ−1γ,S(−k)

∣∣∣ , (8)

where G is a user-selected finite subset of the non-negative real line, S(−k) is the sample co-

variance matrix from the observations outside the kth fold, and S(k) is the sample covariance

matrix from the observations in the kth fold centered by the sample mean of the observations

outside the kth fold. We estimate ∆−1∗ using (7) with its tuning parameter selected by (8) and

S = (X− YηL1)T(X− YηL1)/n. Similarly, we estimate Σ−1∗Y Y using (7) with its tuning param-

eter selected by (8) and S = YTY/n.

4·2. Reduced-rank inverse regression

We propose indirect estimators of β∗ that presupposes that the inverse regression’s coefficient

matrix η∗ is rank-deficient. The following proposition links rank deficiency in η∗ and its estimator

to β∗ and its indirect estimator.

PROPOSITION 3. If Σ∗ is positive definite, then rank(β∗) = rank(η∗). In addition, if Σ−1Y Y

and ∆−1 are positive definite in the indirect estimator β defined by (4), then rank(β) = rank(η).

The proof of this proposition is excluded to save space.

We propose two reduced-rank indirect estimators of β∗ by inserting estimators of η∗,∆−1∗ ,

and Σ∗Y Y in (4). The first estimates Σ∗Y Y with YTY/n and estimates (η∗,∆−1∗ ) with normal

likelihood reduced-rank inverse regression:

(η(r), ∆−1(r)) = arg min(η,Ω)∈Rq×p×Sp+

[n−1tr

(X− Yη)T(X− Yη)Ω

− log det(Ω)

](9)

subject to rank(η) = r,

Page 8: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

337

338

339

340

341

342

343

344

345

346

347

348

349

350

351

352

353

354

355

356

357

358

359

360

361

362

363

364

365

366

367

368

369

370

371

372

373

374

375

376

377

378

379

380

381

382

383

384

8 AARON J. MOLSTAD AND ADAM J. ROTHMAN

where r is selected from 0, . . . ,min(p, q). The solution to (9) is available in closed form

(Reinsel & Velu, 1998).

The second reduced-rank indirect estimator of β∗ estimates η∗ with η(r) defined in (9),

estimates Σ−1∗Y Y with (7) using S = YTY/n, and estimates ∆−1

∗ with (7) using S = (X−

Yη(r))T(X− Yη(r))/n.

The first indirect estimator is likelihood-based and the second indirect estimator exploits spar-

sity in Σ−1∗Y Y and ∆−1

∗ . Neither estimator is defined when min(p, q) > n. In this case, which we

do not address, a regularized reduced-rank estimator of η∗ could be used instead of the estima-

tor defined in (9), e.g., the factor estimation and selection estimator (Yuan et al., 2007) or the

reduced-rank ridge regression estimator (Mukherjee & Zhu, 2011).

5. SIMULATIONS

5·1. Sparse inverse regression simulation

For 200 independent replications, we generated a realization of n independent copies of

(XT, Y T)T, where Y ∼ Nq(0,Σ∗Y Y ) and X | Y = y ∼ Np(ηT∗ y,∆∗). The (i, j)th entry of

Σ∗Y Y was set to ρ|i−j|Y and the (i, j)th entry of ∆∗ was set to ρ|i−j|∆ . We set η∗ = Z A, where

Z had entries independently drawn from N(0, 1), A had entries independently drawn from the

Bernoulli distribution with nonzero probability s∗, and is the element-wise product. This model

is ideal for the example estimator from Section 4·1 because ∆−1∗ and Σ−1

∗Y Y are both sparse. In

the settings we considered, every entry in the corresponding randomly generated β∗ is nonzero

with high probability, but the magnitudes of these entries are small. This motivated us to compare

our indirect estimators of β∗ to direct estimators of β∗ that use penalized least squares.

Page 9: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

385

386

387

388

389

390

391

392

393

394

395

396

397

398

399

400

401

402

403

404

405

406

407

408

409

410

411

412

413

414

415

416

417

418

419

420

421

422

423

424

425

426

427

428

429

430

431

432

Indirect multivariate response linear regression 9

To evaluate performance, we used model error (Breiman & Friedman, 1997; Yuan et al., 2007),

defined as

ME(β, β∗) = tr

(β − β∗)TΣ∗XX(β − β∗). (10)

In each replication, we recorded the observed model error for I1, the indirect estimator proposed

in Section 4·1; IS , the indirect estimator defined by (4) with η defined by (6), ΣY Y = YTY/n,

and ∆ = (X− YηL1)T(X− YηL1)/n;O∆, a part-oracle indirect estimator defined by (4) with η

defined by (6), Σ−1Y Y defined by (7), and ∆−1 = ∆−1

∗ ; O, a part-oracle indirect estimator defined

by (4) with η defined by (6), Σ−1Y Y = Σ−1

∗Y Y , and ∆−1 = ∆−1∗ ; and OY , a part-oracle indirect

estimator defined by (4) with η defined by (6), Σ−1Y Y = Σ−1

∗Y Y , and ∆−1 defined by (7). We

also recorded the observed model error for the ordinary least squares estimator (XTX)−1XTY

when n > p; and the Moore–Penrose least squares estimator X−Y, where X− is the Moore–

Penrose generalized inverse of X when n ≤ p. In addition, we recorded the observed model error

for the estimator formed by q separate univariate ridge regressions, where tuning parameters

were chosen separately; and the multivariate ridge regression estimator, where a single tuning

parameter was chosen.

We selected the tuning parameters for uses of (6) with 5-fold cross-validation, minimizing

validation prediction error on the inverse regression. Tuning parameters for the ridge regression

estimators were selected with 5-fold cross-validation, minimizing validation prediction error on

the forward regression. We selected tuning parameters for uses of (7) with (8). The candidate set

of tuning parameters was

10−8, 10−7.5, . . . , 107.5, 108

.

We display side-by-side boxplots of the model errors from the 200 replications in Fig. 1.

When n = 100, p = 60, q = 60, and s∗ = 0·1, the estimators based on (4) performed well for

both values of ρY that we considered. Our proposed estimator I1 was even competitive with

indirect estimators that used some oracle information. The version of our proposed estimator Is

Page 10: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

433

434

435

436

437

438

439

440

441

442

443

444

445

446

447

448

449

450

451

452

453

454

455

456

457

458

459

460

461

462

463

464

465

466

467

468

469

470

471

472

473

474

475

476

477

478

479

480

10 AARON J. MOLSTAD AND ADAM J. ROTHMAN

(a) ρY = 0·5, ρ∆ = 0·7 (b) ρY = 0·9, ρ∆ = 0·7

Mod

el e

rror

Is I1 OLS L2 R O O∆ OY

0

5

10

15

20

25

30

Mod

el e

rror

Is I1 OLS L2 R O O∆ OY

0

2

4

6

8

10

(c) ρY = 0·5, ρ∆ = 0·7 (d) ρY = 0·9, ρ∆ = 0·7

Mod

el e

rror

I1 MP L2 R O O∆ OY

40

60

80

100

120

140M

odel

err

or

I1 MP L2 R O O∆ OY

10

20

30

40

50

60

Fig. 1. Boxplots of the observed model errors from 200

independent replications when the data generating model

from Section 5·1 was used. In (a) and (b), n = 100, p =

60, q = 60, and s∗ = 0·1. In (c) and (d), n = 50, p = 200,

q = 200, and s∗ = 0·03. The estimator OLS is ordinary

least squares, MP is Moore–Penrose least squares, L2 is q

univariate response ridge regressions with tuning parame-

ters chosen separately, and R is multivariate ridge regres-

sion with one tuning parameter.

that used sample covariance matrices was outperformed by the forward regression estimators.

This suggests that shrinkage estimation of ∆−1∗ and Σ−1

∗Y Y was helpful.

When n = 50, p = 200, q = 200, and s∗ = 0·03, our proposed indirect estimator I1 outper-

formed all three forward regression estimators. The part-oracle method O∆ that used the knowl-

Page 11: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

481

482

483

484

485

486

487

488

489

490

491

492

493

494

495

496

497

498

499

500

501

502

503

504

505

506

507

508

509

510

511

512

513

514

515

516

517

518

519

520

521

522

523

524

525

526

527

528

Indirect multivariate response linear regression 11

edge of ∆−1∗ outperformed the other part-oracle indirect estimator OY , which was slightly better

than I1. Additional results for this model are displayed in the Supplementary Material. In those

results, the performance of I1 relative to the forward regression estimators was similar.

5·2. Non-normal forward regression simulation

For 200 independent replications, we generated n independent copies of (XT, Y T)T

where X ∼ Np (0,Σ∗XX) and Y = βT∗ X + ε. We set ε = Σ

1/2∗E (Z1 − 1, . . . , Zq − 1)T, where

Z1, . . . , Zq are independent copies of an exponential random variable with mean 1. This ensures

that E(ε) = 0 and Cov(ε) = Σ∗E . We indirectly determined the entries of β∗, Σ∗E , and Σ∗XX

by specifying the entries in η∗, ∆−1∗ , and Σ∗Y Y . This required us to use the multivariate nor-

mal model in Section 2·1 even though (XT, Y T)T is not multivariate normal in this simulation.

We set the (i, j)th entry in Σ∗Y Y to ρ|i−j|Y and the (i, j)th entry in ∆∗ to ρ|i−j|∆ . We also set

η∗ = Z A, where Z had entries independently drawn from N(0, 1) and A had entries inde-

pendently drawn from the Bernoulli distribution with nonzero probability s∗. We compared the

performance of the estimators described in Section 5·1 using model error. We selected tuning

parameters in the same way that we did in the simulation described in Section 5·1.

We display side-by-side boxplots of the model errors from the 200 replications in Fig. 2. The

performance of I1 relative to the competitors is similar to how it was in Section 5·1, where

(XT, Y T)T was multivariate normal.

We also performed simulations when (XT, Y T)T had a multivariate elliptical t-distribution.

The results from this simulation are reported in the Supplementary Material. When n = 100,

p = 60, and q = 60, the results from the elliptical t-distribution simulation were similar to the

results here. When n = 50, p = 200, q = 200 and the degrees of freedom of the elliptical t-

distribution was small or the responses had weak marginal correlations, the proposed estimator

Page 12: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

529

530

531

532

533

534

535

536

537

538

539

540

541

542

543

544

545

546

547

548

549

550

551

552

553

554

555

556

557

558

559

560

561

562

563

564

565

566

567

568

569

570

571

572

573

574

575

576

12 AARON J. MOLSTAD AND ADAM J. ROTHMAN

I1 was sometimes outperformed by competitors. These results suggest that our example estimator

may work well for some non-normal data generating models.

(a) ρY = 0·5, ρ∆ = 0·7 (b) ρY = 0·9, ρ∆ = 0·7

Mod

el e

rror

Is I1 OLS L2 R O O∆ OY

0

5

10

15

20

25

30

Mod

el e

rror

Is I1 OLS L2 R O O∆ OY

0

2

4

6

8

10

(c) ρY = 0·5, ρ∆ = 0·7 (d) ρY = 0·9, ρ∆ = 0·7

Mod

el e

rror

I1 MP L2 R O O∆ OY

40

60

80

100

120

140

Mod

el e

rror

I1 MP L2 R O O∆ OY

10

20

30

40

50

60

Fig. 2. Boxplots of the observed model errors from 200

independent replications when the data generating model

from Section 5·2 was used. In (a) and (b), n = 100, p =

60, q = 60, and s∗ = 0·1. In (c) and (d), n = 50, p = 200,

q = 200, and s∗ = 0·03. The estimators are defined in

Section 5·1 and the caption of Fig. 1.

5·3. Reduced-rank inverse regression simulation

For 200 independent replications, we generated a realization of n independent copies of

(XT, Y T)T where Y ∼ Nq(0,Σ∗Y Y ) and X | Y = y ∼ Np(ηT∗ y,∆∗). The (i, j)th entry of

Σ∗Y Y was set to ρ|i−j|Y and the (i, j)th entry of ∆∗ was set to ρ

|i−j|∆ . After specifying

Page 13: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

577

578

579

580

581

582

583

584

585

586

587

588

589

590

591

592

593

594

595

596

597

598

599

600

601

602

603

604

605

606

607

608

609

610

611

612

613

614

615

616

617

618

619

620

621

622

623

624

Indirect multivariate response linear regression 13

r∗ ≤ min(p, q), we set η∗ = PQ, where P ∈ Rq×r∗ had entries independently drawn from

N(0, 1) and Q ∈ Rr∗×p had entries independently drawn from Uniform(−0·25, 0·25) so that

r∗ = rank(η∗) = rank(β∗).

In each replication, we measured the observed model error for IML, the likelihood-based in-

direct first example estimator proposed in Section 4·2; IRR, the second indirect example esti-

mator proposed in Section 4·2, which uses sparse estimators of Σ−1∗Y Y and ∆−1

∗ in (4); OR∆,

a part-oracle indirect estimator defined by (4) with η defined by (9), ∆−1 defined by (7),

and Σ−1Y Y = Σ−1

∗Y Y ; OR, a part-oracle indirect estimator defined by (4) with η defined by (9),

∆−1 = ∆−1∗ , and Σ−1

Y Y = Σ−1∗Y Y ; ORY , a part-oracle indirect estimator defined by (4) with η

defined by (9), ∆−1 = ∆−1∗ , ∆−1 defined by (7), and Σ−1

Y Y defined by (7). We also measured the

observed model error for the direct likelihood-based reduced-rank regression estimator (Izen-

man, 1975; Reinsel & Velu, 1998) and the ordinary least squares estimator.

We selected the rank parameter r for uses of (9) with 5-fold cross-validation, minimizing

validation prediction error on the inverse regression. The rank parameter for the direct likelihood-

based reduced-rank regression estimator was selected with 5-fold cross-validation, minimizing

validation prediction error on the forward regression. We selected tuning parameters for uses of

(7) with (8). The candidate set of tuning parameters was

10−8, 10−7.5, . . . , 107.5, 108

.

We display side-by-side boxplots of the model errors for this reduced-rank inverse regression

simulation in Fig. 3(a) and (b), where we set n = 100, p = 20, q = 20, and r∗ = 4. This choice

of (n, p, q) ensures that IML exists with probability one. When ρY = 0·5, IRR outperformed all

non-oracle competitors. When ρY = 0·9, IRR tended to outperform all non-oracle competitors,

but it performed worse in a small number of replications. Additionally, IRR generally outper-

formed both OR∆ and ORY , which suggests that sparse estimation of ∆−1∗ and Σ−1

∗Y Y was help-

ful. In each setting, IML performed similarly to the direct reduced-rank regression estimator even

Page 14: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

625

626

627

628

629

630

631

632

633

634

635

636

637

638

639

640

641

642

643

644

645

646

647

648

649

650

651

652

653

654

655

656

657

658

659

660

661

662

663

664

665

666

667

668

669

670

671

672

14 AARON J. MOLSTAD AND ADAM J. ROTHMAN

(a) ρY = 0·5, ρ∆ = 0·7 (b) ρY = 0·9, ρ∆ = 0·7M

odel

err

or

IML IRR OLS RR OR OR∆ ORY

0

1

2

3

4

5

Mod

el e

rror

IML IRR OLS RR OR OR∆ ORY

0

1

2

3

4

5

(c) ρX = 0·0, ρE = 0·7 (d) ρX = 0·9, ρE = 0·7

Mod

el e

rror

IML IRR OLS RR OR OR∆ ORY

0

2

4

6

8

10

12

14

Mod

el e

rror

IML IRR OLS RR OR OR∆ ORY

0

1

2

3

4

5

6

7

Fig. 3. Boxplots of the observed model errors from 200

replications when n = 100, p = 20, q = 20, r∗ = 4. In (a)

and (b), the data generating model from Section 5·3 was

used. In (c) and (d), the data generating model from Sec-

tion 5·4 was used. The estimator RR is likelihood-based

reduced-rank forward regression (Izenman, 1975; Reinsel

& Velu, 1998) and OLS is ordinary least squares.

though they are estimating parameters of different conditional distributions. Simulation results

from other data generating models are displayed in the Supplementary Material.

5·4. Reduced-rank forward regression simulation

In this section, we compare the estimators from Section 5·3 using a forward regression data

generating model.

Page 15: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

673

674

675

676

677

678

679

680

681

682

683

684

685

686

687

688

689

690

691

692

693

694

695

696

697

698

699

700

701

702

703

704

705

706

707

708

709

710

711

712

713

714

715

716

717

718

719

720

Indirect multivariate response linear regression 15

For 200 independent replications, we generated a realization of n independent copies of

(XT, Y T)T where X ∼ Np(0,Σ∗XX) and Y | X = x ∼ Nq(βT∗ x,Σ∗E). The (i, j)th entry of

Σ∗XX was set to ρ|i−j|X and the (i, j)th entry of Σ∗E was set to ρ|i−j|E . After specifying r∗ ≤

min(p, q), we set β∗ = ZQ where Z ∈ Rp×r∗ had entries independently drawn from N(0, 1)

and Q ∈ Rr∗×q had entries independently drawn from Uniform(−0·25, 0·25). In this data gen-

erating model, neither ∆−1∗ nor Σ−1

∗Y Y had entries equal to zero.

In each replication, we recorded the observed model error for the estimators described in

Section 5·3. We present boxplots of these model errors from 200 replications with n = 100,

p = 20, q = 20, and r∗ = 4 in Fig. 3 (c) and (d). Both IRR and IML were competitive with

the direct reduced-rank regression estimator. Although neither ∆−1∗ nor Σ−1

∗Y Y were sparse, IRR

generally outperformedORY andOR∆, both of which use some oracle information. These results

demonstrate that using sparse estimators of ∆−1∗ and Σ−1

∗Y Y in (4) may be helpful when neither

is truly sparse.

6. GENOMIC DATA EXAMPLE

We consider a comparative genomic hybridization dataset from Chin et al. (2006) analyzed by

Witten et al. (2009) and Chen et al. (2013). The data are measured gene expression profiles and

DNA copy-number variations for n = 89 subjects with breast cancer. We performed a separate

multivariate response regression analysis for chromosomes 8, 17, and 22. In each analysis, the

q-variate response was DNA copy-number variations and the p-variate predictor was the gene

expression profile. The dimensions for the three analyses were (p, q) = (673, 138), (1161, 87),

and (618, 18).

In the analysis of Chen et al. (2013), estimators that used all p genes significantly outperformed

estimators that used a selected subset of genes. This may indicate that the forward regression co-

Page 16: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

721

722

723

724

725

726

727

728

729

730

731

732

733

734

735

736

737

738

739

740

741

742

743

744

745

746

747

748

749

750

751

752

753

754

755

756

757

758

759

760

761

762

763

764

765

766

767

768

16 AARON J. MOLSTAD AND ADAM J. ROTHMAN

efficient matrix is not sparse. When analyzing similar data, Peng et al. (2010) and Yuan et al.

(2012) focused on modeling the inverse regression, which they assumed to be sparse. This moti-

vated us to apply our indirect estimator that also assumes that the inverse regression is sparse.

In each of 1000 replications, we randomly split the data into training and testing sets of sizes

60 and 29, respectively. Within each replication, we standardized the training dataset predictors

and responses for model fitting and appropriately rescaled predictions. We fit the multivariate

response linear regression model to the training dataset by estimating the regression coefficient

matrix with non-oracle direct and indirect estimators described in Section 5·1. We modified our

proposed estimator I1 because computing the sparse estimates of ∆−1∗ and Σ−1

Y Y took too much

time for small values of their tuning parameters. We instead used I2, which is the same as I1

except that the sparse estimators of ∆−1∗ and Σ−1

Y Y are replaced by the shrinkage estimator defined

by

argminΩ∈Sp+

tr(ΩS)− log det (Ω) + γ∑j,k

|ωjk|2 , (11)

where S = (Y− XηL1)T(Y− XηL1)/n when we estimate ∆−1∗ , and S = YTY/n when we es-

timate Σ−1∗Y Y . Witten & Tibshirani (2009) derived a closed form solution for (11). This shrinkage

estimator of the inverse regression’s error precision matrix was also used in the data example of

Cook et al. (2013). Tuning parameters were selected using the same procedures described in the

simulation studies of Section 5, except the tuning parameter for ∆−1 was chosen to minimize

5-fold cross-validation prediction error on the forward regression after having fixed η and Σ−1Y Y .

We also fit the model using the Moore–Penrose least squares estimator, q separate lasso regres-

sions, the multivariate group lasso estimator of Obozinski et al. (2011), and both ridge regression

estimators described in Section 5.

Tuning parameters for the direct estimators were chosen to minimize 5-fold cross-validation

prediction error on the forward regression. In each replication, we measured the mean squared

Page 17: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

769

770

771

772

773

774

775

776

777

778

779

780

781

782

783

784

785

786

787

788

789

790

791

792

793

794

795

796

797

798

799

800

801

802

803

804

805

806

807

808

809

810

811

812

813

814

815

816

Indirect multivariate response linear regression 17

scaled prediction error which we define as

||(Ytest − Xtestβ)Λ−1||2F29q

,

where Ytest ∈ R29×q is the test dataset response matrix column-centered by the training dataset

response sample mean, Xtest ∈ R29×p is the test dataset predictor matrix column-centered by the

training dataset predictor sample mean, and Λ ∈ Rq×q is a diagonal matrix with the complete

data response marginal standard deviations on its the diagonal. This measure puts predictions on

the same scale for comparison across the q responses.

The mean squared scaled prediction errors are summarized in Table 1. For all three chromo-

somses, the proposed estimator I2 was better than the Moore–Penrose least square estimator, the

null model, q separate lasso regressions, and the group lasso estimator. Although the proposed

estimator I2 performed similarly to both ridge regression estimators, I2 has the advantage of

fitting an interpretable parsimonious inverse regression with an interesting biological interpreta-

tion. Figure 4 displays a heatmap representing how frequently each inverse regression coefficient

was estimated to be nonzero with method I2 in the 1000 replications for the analysis of Chromo-

some 17. The estimated inverse regression coefficient matrices were 3·18%, 4·05%, and 14·7%

nonzero on average for the analyses of Chromosomes 8, 17, and 22 respectively.

7. DISCUSSION

If one has access to the joint distribution of the predictors and responses, then one could use

shrinkage estimators to fit both the forward and inverse regression models. One could then select

the more parsimonious direction, which could be determined by the complexity of the models

recommended by cross validation. If the inverse regression model is more parsimonious, then

our method could be used to improve prediction in the forward direction. Prediction may be the

only goal, in which case the forward and indirect predictions could be combined.

Page 18: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

817

818

819

820

821

822

823

824

825

826

827

828

829

830

831

832

833

834

835

836

837

838

839

840

841

842

843

844

845

846

847

848

849

850

851

852

853

854

855

856

857

858

859

860

861

862

863

864

18 AARON J. MOLSTAD AND ADAM J. ROTHMAN

Table 1. Mean squared scaled prediction error averaged over 1000 replications times

10 and corresponding standard errors times 10.

Chromosome q p I2 NM MP L1 L1/2 L2 R

8 138 673 6·43 10·08 6·79 7·09 7·36 6·47 6·41

(0·029) (0·052) (0·029) (0·033) (0·035) (0·030) (0·030)

17 87 1161 7·83 10·18 8·18 8·62 8·91 8·04 7·94

(0·046) (0·064) (0·046) (0·049) (0·050) (0·050) (0·049)

22 18 618 6·05 10·37 6·67 6·86 6·62 6·15 6·13

(0·043) (0·086) (0·038) (0·052) (0·047) (0·048) (0·049)

I2, the indirect estimator defined in Section 6; MP, the Moore–Penrose least squares estimator; NM,

the null model; L1, q separate lasso regression estimators; L1/2, the multivariate group lasso estimator

of Obozinski et al. (2011); L2, ridge regressions with tuning parameters chosen separately for each

response; R, the multivariate ridge regression estimators with one tuning parameter chosen for all q

responses.

A referee pointed out that it is expensive to compute an indirect estimator in our class when q is

very large because it requires the inversion of a q by q matrix in (4). This referee also mentioned

that our class of indirect estimators is inapplicable when either the predictors or responses are

categorical.

ACKNOWLEDGMENT

We thank Liliana Forzani for an important discussion. We also thank the editor, associate

editor, and referees for helpful comments. This research is supported in part by a grant from the

U.S. National Science Foundation.

Page 19: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

865

866

867

868

869

870

871

872

873

874

875

876

877

878

879

880

881

882

883

884

885

886

887

888

889

890

891

892

893

894

895

896

897

898

899

900

901

902

903

904

905

906

907

908

909

910

911

912

Indirect multivariate response linear regression 19

Genes

CG

H s

pots

20

40

60

80

200 400 600 800 1000

0

200

400

600

800

1000

Fig. 4. A heatmap displaying the number of replications

out of 1000 for which entries in the inverse regression’s

coefficient matrix were estimated to be nonzero by I2 for

Chromosome 17. Black denotes 1000 and white denotes

zero. The genes were sorted by hierarchical clustering.

SUPPLEMENTARY MATERIAL

Supplementary material available at Biometrika online includes additional simulation studies

and the proof of Proposition 2.

APPENDIX

Proof of Proposition 1

Since Σ∗ is positive definite, we apply the partitioned inverse formula to obtain

Σ−1∗ =

Σ∗XX Σ∗XY

ΣT∗XY Σ∗Y Y

−1

=

∆−1∗ −β∗Σ−1

∗E

−η∗∆−1∗ Σ−1

∗E

,

where ∆∗ = Σ∗XX − Σ∗XY Σ−1∗Y Y ΣT

∗XY and Σ∗E = Σ∗Y Y − ΣT∗XY Σ−1

∗XXΣ∗XY . The symmetry of

Σ−1∗ implies that β∗Σ−1

∗E = (η∗∆−1∗ )T so

β∗ = ∆−1∗ ηT∗ Σ∗E . (A1)

Page 20: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

913

914

915

916

917

918

919

920

921

922

923

924

925

926

927

928

929

930

931

932

933

934

935

936

937

938

939

940

941

942

943

944

945

946

947

948

949

950

951

952

953

954

955

956

957

958

959

960

20 AARON J. MOLSTAD AND ADAM J. ROTHMAN

Using the Woodbury identity,

Σ−1∗E = (Σ∗Y Y − ΣT

∗XY Σ−1∗XXΣ∗XY )−1

= Σ−1∗Y Y + Σ−1

∗Y Y ΣT∗XY

(Σ−1

∗XX − Σ∗XY Σ−1∗Y Y ΣT

∗XY

)−1ΣXY Σ−1

∗Y Y

= Σ−1∗Y Y + η∗∆−1

∗ ηT∗ . (A2)

Using the inverse of the expression above in (A1) establishes the result.

REFERENCES

ADRAGNI, K. P. & COOK, R. D. (2009). Sufficient dimension reduction and prediction in regression. Philosophical

Transactions of the Royal Society of London A: Mathematical, Physical and Engineering Sciences 367, 4385–

4405.

BANERJEE, O., EL GHAOUI, L. & D’ASPREMONT, A. (2008). Model selection through sparse maximum likelihood

estimation. Journal of Machine Learning Research 9, 485–516.

BHADRA, A. & MALLICK, B. K. (2013). Joint high-dimensional bayesian variable and covariance selection with an

application to eqtl analysis. Biometrics 69, 447–457.

BREIMAN, L. & FRIEDMAN, J. H. (1997). Predicting multivariate responses in multiple linear regression. Journal

of the Royal Statistical Society: Series B (Statistical Methodology) 59, 3–54.

CHEN, K., DONG, H. & CHAN, K.-S. (2013). Reduced rank regression via adaptive nuclear norm penalization.

Biometrika 100, 902–920.

CHEN, L. & HUANG, J. Z. (2012). Sparse reduced-rank regression for simultaneous dimension reduction and variable

selection. Journal of the American Statistical Association 107, 1533–1545.

CHIN, K., DEVRIES, S., FRIDLYAND, J., SPELLMAN, P. T., ROYDASGUPTA, R., KUO, W.-L., LAPUK, A., NEVE,

R. M., QIAN, Z., RYDER, T. et al. (2006). Genomic and transcriptional aberrations linked to breast cancer

pathophysiologies. Cancer Cell 10, 529–541.

COOK, R. D., FORZANI, L. & ROTHMAN, A. J. (2012). Estimating sufficient reductions of the predictors in abundant

high-dimensional regressions. The Annals of Statistics 40, 353–384.

COOK, R. D., FORZANI, L. & ROTHMAN, A. J. (2013). Prediction in abundant high-dimensional linear regression.

Electronic Journal of Statistics 7, 3059–3088.

COOK, R. D., LI, B. & CHIAROMONTE, F. (2010). Envelope models for parsimonious and efficient multivariate

linear regression (with discussion). Statistica Sinica 20, 927–1010.

Page 21: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

961

962

963

964

965

966

967

968

969

970

971

972

973

974

975

976

977

978

979

980

981

982

983

984

985

986

987

988

989

990

991

992

993

994

995

996

997

998

999

1000

1001

1002

1003

1004

1005

1006

1007

1008

Indirect multivariate response linear regression 21

FRIEDMAN, J. H., HASTIE, T. J. & TIBSHIRANI, R. J. (2008). Sparse inverse covariance estimation with the

graphical lasso. Biostatistics 9, 432–441.

HSIEH, C.-J., SUSTIK, M. A., DHILLON, I. S. & RAVIKUMAR, P. K. (2011). Sparse inverse covariance matrix

estimation using quadratic approximation. In Advances in Neural Information Processing Systems, vol. 24. MIT

Press, Cambridge, MA, pp. 2330–2338.

HUANG, J., LIU, N., POURAHMADI, M. & LIU, L. (2006). Covariance matrix selection and estimation via penalised

normal likelihood. Biometrika 93, 85–98.

IZENMAN, A. J. (1975). Reduced-rank regression for the multivariate linear model. Journal of Multivariate Analysis

5, 248–264.

LAM, C. & FAN, J. (2009). Sparsistency and rates of convergence in large covariance matrices estimation. Annals

of Statistics 37, 4254–4278.

LEE, W. & LIU, Y. (2012). Simultaneous multiple response regression and inverse covariance matrix estimation via

penalized gaussian maximum likelihood. Journal of Multivariate Analysis 111, 241–255.

MUKHERJEE, A. & ZHU, J. (2011). Reduced rank ridge regression and its kernel extensions. Statistical Analysis

and Data Mining 4, 612–622.

OBOZINSKI, G., TASKAR, B. & JORDAN, M. I. (2010). Joint covariate selection and joint subspace selection for

multiple classification problems. Statistics and Computing 20, 231–252.

OBOZINSKI, G., WAINWRIGHT, M. J. & JORDAN, M. I. (2011). Support union recovery in high-dimensional

multivariate regression. The Annals of Statistics 39, 1–47.

PENG, J., ZHU, J., BERGAMASCHI, A., HAN, W., NOH, D.-Y., POLLACK, J. R. & WANG, P. (2010). Regularized

multivariate regression for identifying master predictors with application to integrative genomics study of breast

cancer. The Annals of Applied Statistics 4, 53.

RAVIKUMAR, P., WAINWRIGHT, M. J., RASKUTTI, G. & YU, B. (2011). High-dimensional covariance estimation

by minimizing l1-penalized log-determinant divergence. Electronic Journal of Statistics 5, 935–980.

REINSEL, G. C. & VELU, R. P. (1998). Multivariate Reduced-rank Regression: Theory and Applications. New

York: Springer.

ROTHMAN, A. J., BICKEL, P. J., LEVINA, E. & ZHU, J. (2008). Sparse permutation invariant covariance estimation.

Electronic Journal of Statistics 2, 494–515.

ROTHMAN, A. J., LEVINA, E. & ZHU, J. (2010). Sparse multivariate regression with covariance estimation. Journal

of Computational and Graphical Statistics 19, 947–962.

Page 22: Indirect multivariate response linear regressionusers.stat.umn.edu/~arothman/MolstadRothman-manuscript.pdf · 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162

1009

1010

1011

1012

1013

1014

1015

1016

1017

1018

1019

1020

1021

1022

1023

1024

1025

1026

1027

1028

1029

1030

1031

1032

1033

1034

1035

1036

1037

1038

1039

1040

1041

1042

1043

1044

1045

1046

1047

1048

1049

1050

1051

1052

1053

1054

1055

1056

22 AARON J. MOLSTAD AND ADAM J. ROTHMAN

SU, Z. & COOK, R. D. (2011). Partial envelopes for efficient estimation in multivariate linear regression. Biometrika

98, 133–146.

SU, Z. & COOK, R. D. (2012). Inner envelopes: Efficient estimation in multivariate linear regression. Biometrika

99, 687–702.

SU, Z. & COOK, R. D. (2013). Scaled envelopes: Scale invariant and efficient estimation in multivariate linear

regression. Biometrika 100, 921–938.

TIBSHIRANI, R. J. (1996). Regression shrinkage and selection via the lasso. J. Roy. Statist. Soc., Ser. B 58, 267–288.

TURLACH, B. A., VENABLES, W. N. & WRIGHT, S. J. (2005). Simultaneous variable selection. Technometrics 47,

349–363.

WITTEN, D. M. & TIBSHIRANI, R. J. (2009). Covariance-regularized regression and classification for high dimen-

sional problems. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 71, 615–636.

WITTEN, D. M., TIBSHIRANI, R. J. & HASTIE, T. J. (2009). A penalized matrix decomposition, with applications

to sparse principal components and canonical correlation analysis. Biostatistics 10, 515–534.

YUAN, M. (2008). Efficient computation of l1 regularized estimates in gaussian graphical models. Journal of

Computational and Graphical Statistics 17, 809–826.

YUAN, M., EKICI, A., LU, Z. & MONTEIRO, R. (2007). Dimension reduction and coefficient estimation in multi-

variate linear regression. Journal of the Royal Statistical Society: Series B (Statistical Methodology) 69, 329–346.

YUAN, M. & LIN, Y. (2007). Model selection and estimation in the gaussian graphical model. Biometrika 94, 19–35.

YUAN, Y., CURTIS, C., CALDAS, C. & MARKOWETZ, F. (2012). A sparse regulatory network of copy-number

driven gene expression reveals putative breast cancer oncogenes. IEEE/ACM Transactions on Computational

Biology and Bioinformatics (TCBB) 9, 947–954.